|
import os |
|
import json |
|
import random |
|
import string |
|
import subprocess |
|
import tempfile |
|
import logging |
|
import argparse |
|
from github import Github |
|
from git import Repo |
|
from datasets import load_dataset, Dataset |
|
|
|
|
|
logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s') |
|
logger = logging.getLogger(__name__) |
|
|
|
|
|
GITHUB_TOKEN = os.environ.get("GITHUB_TOKEN") |
|
HF_TOKEN = os.environ.get("HF_TOKEN") |
|
|
|
if not GITHUB_TOKEN: |
|
logger.error("GITHUB_TOKEN environment variable is not set.") |
|
raise ValueError("GITHUB_TOKEN environment variable is not set. Please set it before running the script.") |
|
|
|
if not HF_TOKEN: |
|
logger.error("HF_TOKEN environment variable is not set.") |
|
raise ValueError("HF_TOKEN environment variable is not set. Please set it before running the script.") |
|
|
|
|
|
g = Github(GITHUB_TOKEN) |
|
|
|
def search_top_repos(): |
|
"""Search for top 100 Python repositories with at least 1000 stars and 100 forks.""" |
|
logger.info("Searching for top 100 Python repositories...") |
|
query = "language:python stars:>=1000 forks:>=100" |
|
repos = g.search_repositories(query=query, sort="stars", order="desc") |
|
top_repos = list(repos[:100]) |
|
logger.info(f"Found {len(top_repos)} repositories") |
|
return top_repos |
|
|
|
def clone_repo(repo, tmp_dir): |
|
"""Clone a repository to a temporary directory.""" |
|
logger.info(f"Cloning repository: {repo.full_name}") |
|
repo_dir = os.path.join(tmp_dir, repo.name) |
|
Repo.clone_from(repo.clone_url, repo_dir) |
|
logger.info(f"Repository cloned to {repo_dir}") |
|
return repo_dir |
|
|
|
def run_semgrep(repo_dir): |
|
"""Run Semgrep on the repository and return the JSON output.""" |
|
logger.info(f"Running Semgrep on {repo_dir}") |
|
cmd = f"semgrep scan --config auto --json {repo_dir}" |
|
result = subprocess.run(cmd, shell=True, capture_output=True, text=True) |
|
logger.info("Semgrep scan completed") |
|
return json.loads(result.stdout) |
|
|
|
def extract_vulnerable_files(semgrep_output): |
|
"""Extract files with exactly one vulnerability and their CWE.""" |
|
logger.info("Extracting vulnerable files from Semgrep output") |
|
vulnerable_files = {} |
|
total_vulns = 0 |
|
for result in semgrep_output.get("results", []): |
|
file_path = result.get("path") |
|
cwe = result.get("extra", {}).get("metadata", {}).get("cwe", "Unknown") |
|
|
|
if file_path not in vulnerable_files: |
|
vulnerable_files[file_path] = {"count": 0, "cwe": cwe} |
|
|
|
vulnerable_files[file_path]["count"] += 1 |
|
total_vulns += 1 |
|
|
|
single_vulnerability_files = {file: info["cwe"] for file, info in vulnerable_files.items() if info["count"] == 1} |
|
logger.info(f"Found {total_vulns} total vulnerabilities") |
|
logger.info(f"Found {len(single_vulnerability_files)} files with exactly one vulnerability") |
|
return single_vulnerability_files, total_vulns |
|
|
|
def count_tokens(text): |
|
"""Approximate token count using whitespace splitting.""" |
|
return len(text.split()) |
|
|
|
def generate_random_filename(): |
|
"""Generate a random 6-digit filename with .py extension.""" |
|
return ''.join(random.choices(string.digits, k=6)) + ".py" |
|
|
|
def process_repository(repo, output_file): |
|
"""Process a single repository and append new data items to the output file.""" |
|
logger.info(f"Processing repository: {repo.full_name}") |
|
with tempfile.TemporaryDirectory() as tmp_dir: |
|
repo_dir = clone_repo(repo, tmp_dir) |
|
semgrep_output = run_semgrep(repo_dir) |
|
vulnerable_files, total_vulns = extract_vulnerable_files(semgrep_output) |
|
|
|
items_added = 0 |
|
for file_path, cwe in vulnerable_files.items(): |
|
if items_added >= 3: |
|
logger.info(f"Reached maximum of 3 items for repository {repo.full_name}. Stopping processing.") |
|
break |
|
|
|
full_path = os.path.join(repo_dir, file_path) |
|
logger.info(f"Analyzing file: {file_path}") |
|
with open(full_path, 'r') as f: |
|
source_code = f.read() |
|
|
|
token_count = count_tokens(source_code) |
|
if 512 <= token_count <= 1024: |
|
new_item = { |
|
"source": source_code, |
|
"file_name": generate_random_filename(), |
|
"cwe": cwe |
|
} |
|
|
|
with open(output_file, 'a') as f: |
|
json.dump(new_item, f) |
|
f.write('\n') |
|
items_added += 1 |
|
logger.info(f"Added new item with CWE: {cwe}") |
|
else: |
|
logger.info(f"File skipped: token count ({token_count}) out of range") |
|
|
|
logger.info(f"Processed {repo.full_name}: found {total_vulns} vulnerabilities, added {items_added} new items") |
|
|
|
def preprocess_data(data): |
|
"""Ensure all fields are consistently typed across all items.""" |
|
if not data: |
|
return data |
|
|
|
|
|
list_fields = set() |
|
for item in data: |
|
for key, value in item.items(): |
|
if isinstance(value, list): |
|
list_fields.add(key) |
|
|
|
|
|
for item in data: |
|
for key in list_fields: |
|
if key not in item: |
|
item[key] = [] |
|
elif not isinstance(item[key], list): |
|
item[key] = [item[key]] |
|
|
|
return data |
|
|
|
def merge_and_push_dataset(jsonl_file, new_dataset_name): |
|
"""Push to Hugging Face.""" |
|
logging.info("Starting dataset push process") |
|
|
|
|
|
logging.info("Loading new data from JSONL file") |
|
with open(jsonl_file, 'r') as f: |
|
new_data = [json.loads(line) for line in f] |
|
|
|
logging.info(f"Loaded {len(new_data)} records from JSONL file") |
|
|
|
|
|
logging.info("Preprocessing data") |
|
preprocessed_data = preprocess_data(new_data) |
|
|
|
|
|
logging.info("Creating dataset") |
|
try: |
|
dataset = Dataset.from_list(preprocessed_data) |
|
except Exception as e: |
|
logging.error(f"Error creating dataset: {str(e)}") |
|
|
|
|
|
logging.info(f"Pushing dataset with {len(dataset)} records to Hugging Face") |
|
dataset.push_to_hub(new_dataset_name, private=True, token=HF_TOKEN) |
|
|
|
logging.info("Dataset push process completed") |
|
|
|
def main(): |
|
parser = argparse.ArgumentParser(description="Extend and upload static-analysis-eval dataset") |
|
parser.add_argument("--push_to_dataset", help="Merge and push dataset to specified Hugging Face repository") |
|
args = parser.parse_args() |
|
|
|
if args.push_to_dataset: |
|
|
|
jsonl_file = "static_analysis_eval.jsonl" |
|
merge_and_push_dataset(jsonl_file, args.push_to_dataset) |
|
else: |
|
|
|
output_file = "static_analysis_eval.jsonl" |
|
logger.info(f"Starting dataset extension process. Output file: {output_file}") |
|
|
|
|
|
open(output_file, 'a').close() |
|
|
|
top_repos = search_top_repos() |
|
|
|
for i, repo in enumerate(top_repos, 1): |
|
try: |
|
logger.info(f"Processing repository {i} of {len(top_repos)}: {repo.full_name}") |
|
process_repository(repo, output_file) |
|
except Exception as e: |
|
logger.error(f"Error processing repository {repo.full_name}: {str(e)}", exc_info=True) |
|
|
|
logger.info("Dataset extension process completed") |
|
|
|
if __name__ == "__main__": |
|
main() |