import os import subprocess import json import logging from github import Github import tiktoken from pathlib import Path import shutil # Set up logging logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s') # Initialize GitHub client github_token = os.getenv('GITHUB_TOKEN') if not github_token: raise ValueError("GITHUB_TOKEN environment variable is not set") g = Github(github_token) # Initialize tokenizer tokenizer = tiktoken.get_encoding("cl100k_base") def clone_repo(repo_url, repo_name): tmp_dir = f"/tmp/{repo_name}" subprocess.run(["git", "clone", "--depth", "1", repo_url, tmp_dir], check=True) return tmp_dir def get_repo_content(repo_dir): content = [] for root, dirs, files in os.walk(repo_dir): if "test" in root.lower() or "example" in root.lower(): continue for file in files: if file.endswith('.py'): file_path = os.path.join(root, file) relative_path = os.path.relpath(file_path, repo_dir) with open(file_path, 'r') as f: file_content = f.read() content.append(f"File: {relative_path}\n\n{file_content}\n\n") return "\n".join(content) def count_tokens(text): return len(tokenizer.encode(text)) def process_repo(repo): repo_name = repo.name repo_url = repo.clone_url logging.info(f"Processing repository: {repo_name}") try: tmp_dir = clone_repo(repo_url, repo_name) readme_path = os.path.join(tmp_dir, "README.md") if not os.path.exists(readme_path): logging.info(f"README.md not found in {repo_name}") return None repo_content = get_repo_content(tmp_dir) if count_tokens(repo_content) >= 100000: logging.info(f"Repository {repo_name} content exceeds 100k tokens") return None with open(readme_path, 'r') as f: readme_content = f.read() repo_commit = subprocess.check_output(["git", "rev-parse", "HEAD"], cwd=tmp_dir).decode().strip() return { "repo_name": repo_name, "repo_commit": repo_commit, "repo_content": repo_content, "repo_readme": readme_content } except Exception as e: logging.error(f"Error processing repository {repo_name}: {str(e)}") return None finally: if 'tmp_dir' in locals(): shutil.rmtree(tmp_dir) def load_existing_data(filename): existing_data = {} if os.path.exists(filename): with open(filename, "r") as f: for line in f: item = json.loads(line) existing_data[item['repo_name']] = item return existing_data def save_dataset(filename, dataset, mode='a'): with open(filename, mode) as f: for item in dataset: json.dump(item, f) f.write("\n") def main(): g = Github(os.getenv('GITHUB_TOKEN')) filename = "generate-readme-eval.jsonl" existing_data = load_existing_data(filename) new_dataset = [] updated_count = 0 skipped_count = 0 repos = g.search_repositories(query="language:python stars:>1000 forks:>100", sort="stars", order="desc") for i, repo in enumerate(repos[200:400]): if repo.full_name in existing_data: existing_item = existing_data[repo.full_name] if existing_item['repo_commit'] == repo.get_commits()[0].sha: skipped_count += 1 logging.info(f"Skipped {repo.full_name}: Already processed with same commit") continue else: logging.info(f"Updating {repo.full_name}: Commit changed") updated_count += 1 item = process_repo(repo) if item: new_dataset.append(item) if i % 10 == 0: logging.info(f"Processed {i+1} repositories") # Append new and updated items to the file save_dataset(filename, new_dataset, mode='a') logging.info(f"Dataset updated with {len(new_dataset)} new/updated items") logging.info(f"Skipped {skipped_count} repositories (no changes)") logging.info(f"Updated {updated_count} repositories") if __name__ == "__main__": main()