mirror of
https://github.com/Kludex/awesome-fastapi-projects.git
synced 2025-05-12 20:35:35 +00:00
160 lines
5.2 KiB
Python
160 lines
5.2 KiB
Python
"""The logic for scraping the source graph data processing it."""
|
|
import asyncio
|
|
from typing import NewType
|
|
|
|
import sqlalchemy.dialects.sqlite
|
|
import typer
|
|
from sqlalchemy.ext.asyncio import AsyncSession
|
|
|
|
from app.database import Dependency, Repo, RepoDependency
|
|
from app.dependencies import acquire_dependencies_data_for_repository
|
|
from app.source_graph.client import AsyncSourceGraphSSEClient
|
|
from app.source_graph.mapper import create_or_update_repos_from_source_graph_repos_data
|
|
from app.uow import async_session_uow
|
|
|
|
|
|
async def _create_dependencies_for_repo(session: AsyncSession, repo: Repo) -> None:
|
|
"""
|
|
Create dependencies for a repo.
|
|
|
|
For each parsed dependency, creates a new record in the database, if such a
|
|
dependency does not exist.
|
|
Then, assigns the dependencies to the given repo.
|
|
|
|
:param session: An asynchronous session object
|
|
:param repo: A repo for which to create and assign the dependencies
|
|
"""
|
|
try:
|
|
(
|
|
revision,
|
|
dependencies_create_data,
|
|
) = await acquire_dependencies_data_for_repository(repo)
|
|
except RuntimeError:
|
|
# If the parsing fails, just skip creating the dependencies
|
|
return
|
|
if not dependencies_create_data:
|
|
# If there are no dependencies, just skip creating the dependencies
|
|
return
|
|
# Update the repo with the revision hash
|
|
if repo.last_checked_revision != revision:
|
|
update_repo_statement = (
|
|
sqlalchemy.update(Repo)
|
|
.where(Repo.id == repo.id)
|
|
.values(last_checked_revision=revision)
|
|
)
|
|
await session.execute(update_repo_statement)
|
|
# Create dependencies - on conflict do nothing.
|
|
insert_statement = sqlalchemy.dialects.sqlite.insert(
|
|
Dependency
|
|
).on_conflict_do_nothing(index_elements=[Dependency.name])
|
|
await session.execute(
|
|
insert_statement.returning(Dependency),
|
|
[
|
|
{
|
|
"name": dependency_data.name,
|
|
}
|
|
for dependency_data in dependencies_create_data
|
|
],
|
|
)
|
|
# Re-fetch the dependencies from the database
|
|
dependencies = (
|
|
await session.scalars(
|
|
sqlalchemy.select(Dependency).where(
|
|
Dependency.name.in_(
|
|
[
|
|
dependency_data.name
|
|
for dependency_data in dependencies_create_data
|
|
]
|
|
)
|
|
)
|
|
)
|
|
).all()
|
|
# Add the dependencies to the repo
|
|
repo_dependencies_insert_statement = sqlalchemy.dialects.sqlite.insert(
|
|
RepoDependency
|
|
).on_conflict_do_nothing([RepoDependency.repo_id, RepoDependency.dependency_id])
|
|
await session.execute(
|
|
repo_dependencies_insert_statement,
|
|
[
|
|
{
|
|
"repo_id": repo.id,
|
|
"dependency_id": dependency.id,
|
|
}
|
|
for dependency in dependencies
|
|
],
|
|
)
|
|
|
|
|
|
async def scrape_source_graph_repos() -> None:
|
|
"""
|
|
Iterate over the source graph repos and create or update them in the database.
|
|
|
|
:return: None
|
|
"""
|
|
async with AsyncSourceGraphSSEClient() as sg_client:
|
|
async with async_session_uow() as session:
|
|
async with asyncio.TaskGroup() as tg:
|
|
async for sg_repos_data in sg_client.aiter_fastapi_repos():
|
|
tg.create_task(
|
|
create_or_update_repos_from_source_graph_repos_data(
|
|
session=session,
|
|
source_graph_repos_data=sg_repos_data,
|
|
)
|
|
)
|
|
await session.commit()
|
|
|
|
|
|
DatabaseRepoId = NewType("DatabaseRepoId", int)
|
|
|
|
|
|
async def parse_dependencies_for_repo(
|
|
semaphore: asyncio.Semaphore, repo_id: DatabaseRepoId
|
|
) -> None:
|
|
"""
|
|
Parse the dependencies for a given repo and create them in the database.
|
|
|
|
:param semaphore: A semaphore to limit the number of concurrent requests
|
|
:param repo_id: The id of the repo for which to parse the dependencies
|
|
:return: None
|
|
"""
|
|
async with async_session_uow() as session, semaphore:
|
|
repo = (
|
|
await session.scalars(sqlalchemy.select(Repo).where(Repo.id == repo_id))
|
|
).one()
|
|
# Create the dependencies for the repo
|
|
await _create_dependencies_for_repo(session=session, repo=repo)
|
|
await session.commit()
|
|
|
|
|
|
async def parse_dependencies_for_repos() -> None:
|
|
"""
|
|
Parse the dependencies for all the repos in the database.
|
|
|
|
:return: None.
|
|
"""
|
|
async with async_session_uow() as session:
|
|
repo_ids = (await session.scalars(sqlalchemy.select(Repo.id))).all()
|
|
semaphore = asyncio.Semaphore(10)
|
|
async with asyncio.TaskGroup() as tg:
|
|
for repo_id in repo_ids:
|
|
tg.create_task(
|
|
parse_dependencies_for_repo(
|
|
semaphore=semaphore, repo_id=DatabaseRepoId(repo_id)
|
|
)
|
|
)
|
|
|
|
|
|
def main() -> None:
|
|
"""
|
|
Scrape the FastAPI-related repositories utilizing the source graph API.
|
|
|
|
For each scraped repository, parse the dependencies and create them in the database.
|
|
:return:
|
|
"""
|
|
asyncio.run(scrape_source_graph_repos())
|
|
asyncio.run(parse_dependencies_for_repos())
|
|
|
|
|
|
if __name__ == "__main__":
|
|
typer.run(main)
|