diff --git a/TalentLLM-main/.gitignore b/TalentLLM-main/.gitignore new file mode 100644 index 0000000000000000000000000000000000000000..c0887b35df24ad8e1ccf4dccd3579917a1741bfb --- /dev/null +++ b/TalentLLM-main/.gitignore @@ -0,0 +1,2 @@ +service_creds.json +.env \ No newline at end of file diff --git a/TalentLLM-main/.vscode/launch.json b/TalentLLM-main/.vscode/launch.json new file mode 100644 index 0000000000000000000000000000000000000000..325838697c495b094e757fae128730f23fb8b083 --- /dev/null +++ b/TalentLLM-main/.vscode/launch.json @@ -0,0 +1,16 @@ +{ + "version": "0.2.0", + "configurations": [ + { + "name": "Python: Current File (pdb)", + "type": "python", + "request": "launch", + "program": "${file}", + "console": "integratedTerminal", + "python": "python3.11", + "cwd": "${workspaceFolder}", + "env": {}, + "internalConsoleOptions": "neverOpen" + } + ] +} diff --git a/TalentLLM-main/Candidate.py b/TalentLLM-main/Candidate.py new file mode 100644 index 0000000000000000000000000000000000000000..0850457b412c6309a43daaa7e553eb019b1486bb --- /dev/null +++ b/TalentLLM-main/Candidate.py @@ -0,0 +1,103 @@ +from typing import Optional +import gdown +import os +from datetime import datetime # Importing the datetime class directly +from gh import getBasicReport +from mathpix import extract_text +from pathlib import Path +class JobCandidate: + def __init__(self, data: list): + self.timestamp = datetime.strptime(data[0], "%m/%d/%Y %H:%M:%S") + self.name = data[1] + self.email = data[2] + self.resume_link = data[3] + self.resume_text= self.parse_resume() + self.cover_letter = data[4] + self.linkedin = data[5] + self.github_link = data[6] + self.github_text= self.parse_gh() + self.personal_website_link = data[7] + self.visa_sponsorship = data[8] + self.disability_status = data[9] + self.ethnic_background = data[10] + self.gender = data[11] + self.military_service = data[12] + + def __str__(self): + return (f"Job Candidate: {self.name}\n" + f"Applied on: {self.timestamp}\n" + f"Email: {self.email}\n" + f"Resume {self.resume_text}\n" + f"Personal Website: {self.personal_website_link}\n" + f"Visa Sponsorship: {self.visa_sponsorship}\n" + f"Disability Status: {self.disability_status}\n" + f"Ethnic Background: {self.ethnic_background}\n" + f"Gender: {self.gender}\n" + f"Military Service: {self.military_service}") + + def parse_resume(self): + id = self.resume_link.split('=')[-1] + pdf_dir = os.path.join(os.getcwd(), "resume_pdfs") + mmd_dir = os.path.join(os.getcwd(), "resume_mmds") + + # Ensure the directories exist + if not os.path.exists(pdf_dir): + os.makedirs(pdf_dir) + if not os.path.exists(mmd_dir): + os.makedirs(mmd_dir) + + pdf_path = os.path.join(pdf_dir, f"{self.email}.pdf") + mmd_path = os.path.join(mmd_dir, f"{self.email}.pdf.mmd") + + try: + # Check if the parsed text already exists + if os.path.exists(mmd_path): + with open(mmd_path, "r") as f: + return f.read() + else: + # Download the PDF + gdown.download(id=id, quiet=True, use_cookies=False, output=pdf_path) + + # Check if the download was successful + if os.path.exists(pdf_path): + t = extract_text(pdf_path) + preproccessed = t.replace(self.name, "applicant") + preprocessed = preproccessed.replace(self.name.split(" ")[0], "applicant") + return preprocessed + else: + return "Failed to download the PDF." + except Exception as e: + return str(e) + + + + def parse_gh(self): + username = self.github_link.replace("https://github.com/", "").replace("github.com", "").replace("/", "") + + summary="" + if username: + file_path = Path(os.getcwd()) / "gh_cache" / f"{username}.md" + if not file_path.exists(): + summary = str(getBasicReport(username)) + # Write the summary to the file + file_path.write_text(summary) + else: + summary = open(file_path,"r").read() + return summary + else: + return "" + def parse_portfolio(self): + pass + + + + + + + def __lt__(self, other): + if not isinstance(other, JobCandidate): + return NotImplemented + return self.timestamp < other.timestamp + + def __eq__(self, other): + return False \ No newline at end of file diff --git a/TalentLLM-main/__pycache__/Candidate.cpython-311.pyc b/TalentLLM-main/__pycache__/Candidate.cpython-311.pyc new file mode 100644 index 0000000000000000000000000000000000000000..b2e9e35e5337406c214dd168e8c92f725a520716 Binary files /dev/null and b/TalentLLM-main/__pycache__/Candidate.cpython-311.pyc differ diff --git a/TalentLLM-main/__pycache__/claude_compator.cpython-311.pyc b/TalentLLM-main/__pycache__/claude_compator.cpython-311.pyc new file mode 100644 index 0000000000000000000000000000000000000000..d9739372074a5fb1f0ec3ce6df5d25cb6ad740c2 Binary files /dev/null and b/TalentLLM-main/__pycache__/claude_compator.cpython-311.pyc differ diff --git a/TalentLLM-main/__pycache__/compator.cpython-311.pyc b/TalentLLM-main/__pycache__/compator.cpython-311.pyc new file mode 100644 index 0000000000000000000000000000000000000000..fcbc5e3688543f214d3b429949a4315da882abe8 Binary files /dev/null and b/TalentLLM-main/__pycache__/compator.cpython-311.pyc differ diff --git a/TalentLLM-main/__pycache__/gh.cpython-311.pyc b/TalentLLM-main/__pycache__/gh.cpython-311.pyc new file mode 100644 index 0000000000000000000000000000000000000000..d2c845f88691fcf22110c53e50efa4c8aae0fd09 Binary files /dev/null and b/TalentLLM-main/__pycache__/gh.cpython-311.pyc differ diff --git a/TalentLLM-main/__pycache__/mathpix.cpython-311.pyc b/TalentLLM-main/__pycache__/mathpix.cpython-311.pyc new file mode 100644 index 0000000000000000000000000000000000000000..f0799404ee67a2e7e78bc29f519be7d81b382378 Binary files /dev/null and b/TalentLLM-main/__pycache__/mathpix.cpython-311.pyc differ diff --git a/TalentLLM-main/__pycache__/results.cpython-311.pyc b/TalentLLM-main/__pycache__/results.cpython-311.pyc new file mode 100644 index 0000000000000000000000000000000000000000..a9ac51ed5d7c6cb6bbc6885e70958b36d61343c3 Binary files /dev/null and b/TalentLLM-main/__pycache__/results.cpython-311.pyc differ diff --git a/TalentLLM-main/__pycache__/resume_conversation.cpython-311.pyc b/TalentLLM-main/__pycache__/resume_conversation.cpython-311.pyc new file mode 100644 index 0000000000000000000000000000000000000000..b980ab5a878a5d3f45b31995d453f470231f2e84 Binary files /dev/null and b/TalentLLM-main/__pycache__/resume_conversation.cpython-311.pyc differ diff --git a/TalentLLM-main/__pycache__/resume_conversation_interactive.cpython-311.pyc b/TalentLLM-main/__pycache__/resume_conversation_interactive.cpython-311.pyc new file mode 100644 index 0000000000000000000000000000000000000000..6e5dd4dd3b8d05714752cfe48957661a8120f504 Binary files /dev/null and b/TalentLLM-main/__pycache__/resume_conversation_interactive.cpython-311.pyc differ diff --git a/TalentLLM-main/__pycache__/set_envs.cpython-311.pyc b/TalentLLM-main/__pycache__/set_envs.cpython-311.pyc new file mode 100644 index 0000000000000000000000000000000000000000..b37ea736847941feef58381eb0517954ed138160 Binary files /dev/null and b/TalentLLM-main/__pycache__/set_envs.cpython-311.pyc differ diff --git a/TalentLLM-main/comparisons.json b/TalentLLM-main/comparisons.json new file mode 100644 index 0000000000000000000000000000000000000000..f0ed001805c40bf3216afa1db8ab93c9c0338a9d --- /dev/null +++ b/TalentLLM-main/comparisons.json @@ -0,0 +1 @@ +{"danikhan632@gmail.com#tawfiqaliu@gmail.com#0": -1, "tawfiqaliu@gmail.com#danikhan632@gmail.com#0": 1, "danikhan632@gmail.com#jamesryzhkov@gmail.com#0": 0, "jamesryzhkov@gmail.com#danikhan632@gmail.com#0": 0, "danikhan632@gmail.com#tawfiqfm@gmail.com#0": -1, "tawfiqfm@gmail.com#danikhan632@gmail.com#0": 1, "tawfiqaliu@gmail.com#jamesryzhkov@gmail.com#0": -1, "jamesryzhkov@gmail.com#tawfiqaliu@gmail.com#0": 1, "tawfiqaliu@gmail.com#tawfiqfm@gmail.com#0": 0, "tawfiqfm@gmail.com#tawfiqaliu@gmail.com#0": 0, "jamesryzhkov@gmail.com#tawfiqfm@gmail.com#0": 1, "tawfiqfm@gmail.com#jamesryzhkov@gmail.com#0": -1, "danikhan632@gmail.com#smufti3@gatech.edu#0": 0, "smufti3@gatech.edu#danikhan632@gmail.com#0": 0, "jamesryzhkov@gmail.com#smufti3@gatech.edu#0": 0, "smufti3@gatech.edu#jamesryzhkov@gmail.com#0": 0, "tawfiqaliu@gmail.com#smufti3@gatech.edu#0": -1, "smufti3@gatech.edu#tawfiqaliu@gmail.com#0": 1, "tawfiqfm@gmail.com#smufti3@gatech.edu#0": 0, "smufti3@gatech.edu#tawfiqfm@gmail.com#0": 0, "tawfiqfm@gmail.com#dan.hus120@gmail.com#0": 1, "dan.hus120@gmail.com#tawfiqfm@gmail.com#0": -1, "smufti3@gatech.edu#dan.hus120@gmail.com#0": 1, "dan.hus120@gmail.com#smufti3@gatech.edu#0": -1, "dan.hus120@gmail.com#danikhan632@gmail.com#0": 0, "danikhan632@gmail.com#dan.hus120@gmail.com#0": 0, "dan.hus120@gmail.com#jamesryzhkov@gmail.com#0": 0, "jamesryzhkov@gmail.com#dan.hus120@gmail.com#0": 0, "dan.hus120@gmail.com#tawfiqaliu@gmail.com#0": 0, "tawfiqaliu@gmail.com#dan.hus120@gmail.com#0": 0} \ No newline at end of file diff --git a/TalentLLM-main/compator.py b/TalentLLM-main/compator.py new file mode 100644 index 0000000000000000000000000000000000000000..48d2f15ce44e9c943e0e082ad8214ee4e0e32d2c --- /dev/null +++ b/TalentLLM-main/compator.py @@ -0,0 +1,234 @@ +import openai; +import json, os,sys +from dotenv import load_dotenv +load_dotenv() +# openai.api_key = os.environ.get("OPENAI_API_KEY") +import openai; openai.api_key = "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" +from Candidate import JobCandidate + + +import litellm +from litellm import completion + + +import xml.etree.ElementTree as ET + + + + + + + +def printc(obj, color="cyan"): + color_code = { + "black": "30", "red": "31", "green": "32", "yellow": "33", + "blue": "34", "magenta": "35", "cyan": "36", "white": "37" + } + colored_text = f"\033[{color_code[color]}m{obj}\033[0m" if color in color_code else obj + print(colored_text) + + + +LLM=os.environ.get("COMPARATOR_LLM","chat-bison") +# LLM=os.environ.get("COMPARATOR_LLM","gpt-3.5-turbo-1106") +def getContent(candidateA, candidateB) -> str: + return ( + "Given the following two candidates, choose between the two. Here is the rubric: " + + get_rubric() + + "Candidate A: " + + "\nRESUME:\n" +candidateA.resume_text+"\nEND Resume\n" + + "\nGITHUB:\n" +candidateA.github_text+"\nEND GITHUB\n" + + " END OF Candidate A" + + "\n\nCandidate B: " + + "\nRESUME:\n" +candidateB.resume_text+"\nEND Resume\n" + + "\nGITHUB:\n" +candidateB.github_text+"\nEND GITHUB\n" + + " END OF Candidate B" + + ) + + + +def google_compare_resumes(content:str, nameA="", nameB=""): + choice =0 + messages=[ + {"role": "user", "content": "You are an LLM recrutier who will choose between two candidates based on an provided rubric"}, + {"role": "user", "content": + """ + You are an LLM recrutier who will choose between two candidates based on an provided rubric, + you will only use bullet point and broken english instead of proper english to be more concise + """ + }, + {"role": "assistant", "content": + """ + I can assist you in evaluating two candidates based on a provided rubric. + Provide me with the rubric or the criteria you'd like to use for the evaluation, + and I'll help you assess the candidates accordingly and explain myself in less that 50 words + """ + }, + {"role": "user", "content": content} + ] + + response =completion(model=LLM, messages=messages,max_tokens=170,) + printc(response["choices"][0]["message"],'red') + + messages=[ + {"role": "assistant","content":str(response["choices"][0]["message"])}, + {"role": "user","content":"okay so now with just a single token select A or B, choice letter goes here"} + ] + retries=3 + while retries >0: + response =completion(model=LLM, messages=messages,max_tokens=5,temperature=0.01) + # printc(response,'cyan') + html=''.join(str(response["choices"][0]["message"]['content']).split()) + if "" in html: + xml_content = f'{html}' + root = ET.fromstring(xml_content) + select_element = root.find('select') + letter = str(select_element.text) + else: + letter = str(html)[0] + + + if letter == 'A': + printc(nameA+" wins over "+nameB,"cyan") + return -1 + elif letter == 'B': + printc(nameB+" wins over "+nameA,"green") + return 1 + + + retries-=1 + + + + return choice + + +def compare_resumes(content:str, nameA="", nameB=""): + retries = 3 + choice = 0 + + while retries > 0: + try: + response = openai.ChatCompletion.create( + model='gpt-4-0613', + messages=[ + {"role": "user", "content": + """ + You are an LLM recrutier who will choose between two candidates based on an provided rubric, + you will only use bullet point and broken english instead of proper english to be more concise in your justification + You will also provide args for selectCandidate + """ + }, + {"role": "assistant", "content": + """ + I can assist you in evaluating two candidates based on a provided rubric. + Provide me with the rubric or the criteria you'd like to use for the evaluation, + and I'll help you assess the candidates accordingly and explain myself conscisely and will + provide args for selectCandidate + """ + }, + {"role": "user", "content": content} + + ], + functions=[ + { + "name": "selectCanidate", + "description": "choose between the two canidates", + "parameters": { + "type": "object", + "properties": { + "choice_num": { + "type": "integer", + "description": "1 for Candidate A is the best fit, 2 for Candidate B is the best fit", + "required": ["choice_num"], + }, + "justifcation": { + "type": "string", + "description": "justifcation for why you chose the candidate", + "required": ["justifcation"], + }, + } + }, + } + ], + function_call="auto", + ) + + message = response["choices"][0]["message"] + + if message.get("function_call"): + function_name = message["function_call"]["name"] + function_args = json.loads(message["function_call"]["arguments"]) + choice = (int(function_args["choice_num"])) + + if function_name == "selectCanidate": + if choice == 1: + choice = -1 + printc(nameA+" wins over "+nameB, "cyan") + elif choice == 2: + choice = 1 + printc(nameB+" wins over "+nameA, "green") + + printc(function_args["justifcation"], "yellow") + + break # Break the loop if everything went well + + except Exception as e: + printc("Error: " + str(e), "red") + retries -= 1 + if retries == 0: + printc("Maximum retries reached.", "red") + return 0 # Or any other default value or error indicator + + return choice + + + + +def get_rubric(): + text = open("rubric.txt","r").read() + return "\nRubric:\n" +str(text)+"\nEND Rubric\n" + + + + + +def comp(candidateA:JobCandidate, candidateB:JobCandidate, rub_id:int=0 ) -> int: + comp_table= json.load(open("comparisons.json","r")) + tag= (candidateA.email+"#"+candidateB.email+"#"+str(rub_id)) + inv_tag= (candidateB.email+"#"+candidateA.email+"#"+str(rub_id)) + if tag in comp_table: + if comp_table[tag]==1: + printc(candidateA.name+" wins over "+candidateB.name,"magenta") + elif comp_table[tag]==-1: + printc(candidateB.name+" wins over "+candidateA.name,"magenta") + + return comp_table[tag] + elif inv_tag in comp_table: + if comp_table[inv_tag]==1: + printc(candidateA.name+" wins over "+candidateB.name,"magenta") + elif comp_table[inv_tag]==-1: + printc(candidateB.name+" wins over "+candidateA.name,"magenta") + else: + choice = compare_resumes(getContent(candidateA, candidateB), candidateA.name, candidateB.name) + comp_table[tag]=choice + comp_table[inv_tag]=choice*-1 + + json.dump(comp_table, open("comparisons.json","w")) + return choice + + +def compute_scores(candidates): + scores = {candidate.email: 0 for candidate in candidates} + for i, candidateA in enumerate(candidates): + for candidateB in candidates[i+1:]: + result = comp(candidateA, candidateB) + scores[candidateA.email] += result + scores[candidateB.email] -= result + print(scores) + return scores + +def bubble_sort(candidates: list) -> list: + scores = compute_scores(candidates) + return sorted(candidates, key=lambda x: scores[x.email]) \ No newline at end of file diff --git a/TalentLLM-main/daniyal.txt b/TalentLLM-main/daniyal.txt new file mode 100644 index 0000000000000000000000000000000000000000..52857ed4185814e1b3d98ca2590e99015eb3fd3f --- /dev/null +++ b/TalentLLM-main/daniyal.txt @@ -0,0 +1,56 @@ +Daniyal - Software Engineer +{ +DH +} +Hey there, I'm +Daniyal Hussain +Digital Craftsman +- a software engineer, data scientist, full-stack dev, and more. +📚 +Graduated with a B.A. in +Computer Science +& +Data Science +@ +NYU +. +🚀 +Looking for fulltime roles! +Github +LinkedIn +Email +⚡ About Me +I'm a computer scientist based in NYC with an avid interest in interdisciplinary learning. My mindset is perfectly embodied by this quote from Naval Ravikant: +"Learn to sell, learn to build - if you can do both, you will be unstoppable." +My specialties lie in developing desktop applications, fullstack websites, machine learning, and making it all visually stunning. +I am currently working on Resyzer - a SaaS windows desktop application which takes window management to the next level. +Projects +See all projects → +Here are a couple of the things I've worked on. +See all projects → +NYU Course Tracker +JavaScript +Puppeteer +Utilises puppeteer to scrape the NYU public course search via a headless browser for the desired class and check its status. Manipulates user agent randomly to avoid detection. Sends an alert via telegram when the class status changes. +KeyboardHookLite +C# +win32 +A lightweight low-level global windows keyboard hook compatible with modern UI frameworks (WPF, MAUI, WinUI 3, etc.) +React Blackjack +React +mongoDB +Express +Node.js +Passport +A blackjack app built with react. Supports user authentication. Your balance and game history are preserved. Implemented in react, node.js, passport, mongoDB. +Get In Touch! +LinkedIn +Email +Resume +Built with +Next.js +& +Chakra UI +. Hosted on +Vercel +. \ No newline at end of file diff --git a/TalentLLM-main/gh.py b/TalentLLM-main/gh.py new file mode 100644 index 0000000000000000000000000000000000000000..f71ea41a2a36305c06e4449aeab8e4a2f5b4777f --- /dev/null +++ b/TalentLLM-main/gh.py @@ -0,0 +1,139 @@ +from github import Github,Auth +from github.Repository import Repository +from typing import List +import requests +import os, json, datetime, re,sys +from bs4 import BeautifulSoup +import openai +from dotenv import load_dotenv +load_dotenv() +import litellm +from litellm import completion +os.environ['AWS_ACCESS_KEY_ID']="AKIAWE73I2DYVFLFNLHS" +os.environ['AWS_REGION_NAME']="us-east-1" +os.environ['AWS_REGION']="us-east-1" +os.environ['AWS_SECRET_ACCESS_KEY']="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" +os.environ['PALM_API_KEY']='AIzaSyBr-t20IcF2T1xItAnlyYuQ50Ctu6Y0y4I' +os.environ["VERTEXAI_PROJECT"] = "data-axe-386317" +os.environ['GOOGLE_APPLICATION_CREDENTIALS'] = 'service_creds.json' + +def printc(obj, color="cyan"): + color_code = { + "black": "30", "red": "31", "green": "32", "yellow": "33", + "blue": "34", "magenta": "35", "cyan": "36", "white": "37" + } + colored_text = f"\033[{color_code[color]}m{obj}\033[0m" if color in color_code else obj + print(colored_text) + + +auth = Auth.Token(os.environ.get('GH_KEY', 'default')) +g = Github(auth=auth) + + +def remove_html_and_urls(markdown_text): + no_html = re.sub(r'<[^>]+>', '', markdown_text) + pattern_urls = r'http[s]?://(?:[a-zA-Z]|[0-9]|[$-_@.&+]|[!*\'(),]|(?:%[0-9a-fA-F][0-9a-fA-F]))+' + no_html_no_urls = re.sub(pattern_urls, '', no_html) + + return no_html_no_urls + + + + +def getGithubPinned(username: str)-> List[str]: + repos = [] + today = datetime.datetime.now() + day_1 = today.replace(day=1) + start_date, end_date = day_1.strftime("%Y-%m-%d"), today.strftime("%Y-%m-%d") + + url = f"https://github.com/{username}?tab=overview&from={start_date}&to={end_date}" + + response = requests.get(url) + + if response.status_code == 200: + soup = BeautifulSoup(response.text, 'html.parser') + pinned_items = soup.find_all('div', class_='pinned-item-list-item-content') + + repos = [] + for item in pinned_items: + repo_name = item.find('span', class_='repo').text.strip() + repos.append(repo_name) + else: + print(f"Failed to get pinned repos for {username}") + + return repos + + + + +def get_repositories(username: str)->List[Repository]: + user = g.get_user(username) + all_repos = [repo for repo in user.get_repos()] + repo_dict = {repo.name: repo for repo in all_repos} + pinned_repo_names = getGithubPinned(username) + pinned_repos = [] + for name in pinned_repo_names: + if name in repo_dict: + pinned_repos.append(repo_dict.pop(name)) + sorted_repos = sorted(repo_dict.values(), key=lambda x: x.stargazers_count, reverse=True) + final_repo_list = pinned_repos + sorted_repos + + return final_repo_list + + + + +def getBasicReport(username: str): + try: + user_repos = get_repositories(username)[:8] + summaries=[] + + + for repo in user_repos: + + try: + content = "" + content+="\nNAME: "+str(repo.full_name)+"\nSTARS: "+str(repo.stargazers_count)+"\nReadme: \n" + files = repo.get_contents("") + md_files = [file for file in files if file.name.endswith('.md')] + + + md_file_content = repo.get_contents(md_files[0].path).decoded_content.decode() + + content+= str(remove_html_and_urls(str(md_file_content))) + + + + messages=[ + {"role": "user", "content": "I want you to summarize this repository and summarize the skills gained with this repository "}, + {"role": "assistant", "content": + """ + Sure, I can help with that! Please provide me with the details for the repo and I'll be able to summarize it and outline the skills that can be gained from it. + Additonally I will grade the techinal complexity with it. I will also greatly take into consideration the Number of stars. Furthermore I Will use broken english to ensure + my statements are as short and concise as possible + """ + }, + {"role": "user", "content": content} + ] + + response =completion(model="anthropic.claude-instant-v1", messages=messages,max_tokens=150,temperature=1.0) + summaries.append(response["choices"][0]["message"]['content']) + + except: + continue + + + # message = completion(model="anthropic.claude-instant-v1", messages=messages) + printc(summaries,'cyan') + + + + + return summaries + except: + return "" + + + + + diff --git a/TalentLLM-main/gh_cache/AzeezIsh.md b/TalentLLM-main/gh_cache/AzeezIsh.md new file mode 100644 index 0000000000000000000000000000000000000000..149f241047a8506fa744b2b27df5b2a2cbade2f3 --- /dev/null +++ b/TalentLLM-main/gh_cache/AzeezIsh.md @@ -0,0 +1 @@ +[' Here is a summary of the AzeezIsh/housing_app repository:\n\n- This is a housing application project that utilizes Figma for frontend design and React for development. \n\n- It is primarily a frontend focused project with minimal backend functionality, as data sources are coming from cleaned datasets rather than a custom backend API.\n\nSkills that can be gained from working on this repository include:\n\n- UI/UX design using Figma for prototyping interfaces \n\n- Frontend development with React, including components, states, props, etc. \n\n- Integrating data from external sources into a React application\n\n- Basic frontend architecture and application structure without backend dependencies\n\nThe technical complexity of this project would be'] \ No newline at end of file diff --git a/TalentLLM-main/gh_cache/PramaYudhistira.md b/TalentLLM-main/gh_cache/PramaYudhistira.md new file mode 100644 index 0000000000000000000000000000000000000000..78a9150dc22f0ca03f39062212b94f1d73f5b688 --- /dev/null +++ b/TalentLLM-main/gh_cache/PramaYudhistira.md @@ -0,0 +1 @@ +[" Here is the summary:\n\nName: PramaYudhistira/sh-web\nStars: 1\n\nThis project aims to build a web app called SideHustle that helps college students earn money by posting their skills and services that other students on campus may need. It focuses on services students can provide locally within a campus since many don't have cars. \n\nThe unique value is that it is similar to Craigslist but tailored specifically for college students. \n\nThe tech stack includes ASP.NET for the backend with Entity Framework and plans to use Angular for the frontend UI initially. There is also a note about potentially building a mobile version with React.\n\nSkills gained would include experience with:\n-", ' Here is a summary of the PramaYudhistira/urc-software repository:\n\nThis repository contains the software for the RoboJackets URC (Urban Reconnaissance and Capabilities) robot. It is organized into various ROS packages that enable capabilities like navigation, perception, manipulation, etc. \n\nSome of the key skills that can be gained from working with this repository include:\n\n- ROS system design and organization into packages\n- Robot perception using sensors like cameras \n- SLAM and navigation for autonomous mobility\n- Robot arm kinematics and manipulation\n- ROS message definitions and inter-package communication\n- Integration of robot hardware drivers\n- ROS simulation using Gazebo\n- Robot teleoperation through web', ' Here is a summary of the PramaYudhistira/software-training repository:\n\nRepo contain resource for software training program. Include syllabus and setup instruction. Syllabus outline what be cover in each session of training. Setup folder provide direction for student to prepare their computer for class. \n\nSkill gain from this repo include:\n\n- Learning software development fundamentals by following syllabus. Give overview of topics.\n\n- Setting up programming environment by following setup instructions. Gain skill of installing tools need for development. \n\nTechnical complexity is low. Mainly text documentation for syllabus and setup. \n\nSmall repo just starting. Not many star since new. But provide base for organizing software training course. Student and', ' Here is a summary of the PramaYudhistira/CS2340 repository:\n\nRepository contains materials for CS2340 course. No description provided in README. \n\nSkills gained:\n\n- Programming fundamentals as it is an introductory CS course. Likely covers basics of programming with a language like Python or Java.\n\n- Algorithm design and analysis. Course likely teaches techniques for solving problems systematically and analyzing time/space complexity.\n\n- Data structures. Repository may include implementations of common data structures like lists, stacks, queues, trees, hashes. \n\nTechnical complexity: Beginner. As an intro CS course, focuses on basic programming concepts and data structures. \n\nWith no other details provided', ' Here is a summary of the PramaYudhistira/VIPTraining repository:\n\nRepository contains training material for V.I.P training. Repository has 0 stars currently. \n\nSkills that can be gained:\n\n- Learn skills around virtual instructor-led training. Ability to create and deliver online training content.\n\n- Training design skills. Ability to structure training modules, create lesson plans and map learning outcomes. \n\n- Online facilitation skills. Skills around engaging learners, handling questions and ensuring knowledge transfer in a virtual environment. \n\n- Training delivery skills. Skills around using tools like video conferencing platforms, presenting content and handling technical glitches during delivery.\n\nTechnical complexity is likely moderate'] \ No newline at end of file diff --git a/TalentLLM-main/gh_cache/Saad-Mufti.md b/TalentLLM-main/gh_cache/Saad-Mufti.md new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/TalentLLM-main/gh_cache/Serhan-Asad.md b/TalentLLM-main/gh_cache/Serhan-Asad.md new file mode 100644 index 0000000000000000000000000000000000000000..b21bb5ae3b7ddea2c3bb1ab44b26d191d862e610 --- /dev/null +++ b/TalentLLM-main/gh_cache/Serhan-Asad.md @@ -0,0 +1 @@ +[' Here is a summary of the Serhan-Asad/Beginner- repository:\n\n- The repository appears to be tracking a 30 month internship as the readme simply states "30 month internship". \n\n- With no other context provided in the readme or any files in the repository, it\'s difficult to determine specifics about the type of internship, skills being learned, projects worked on etc.\n\n- However, some skills that could potentially be gained from a 30 month internship include:\n - Hands-on work experience in the field/industry of the internship\n - Learning technical or soft skills applicable to the role/company\n - Project management experience by potentially working on smaller scoped projects\n - Time management and accountability by', " Here is a summary of the Serhan-Asad/jpmc-task-1 repository:\n\nThis repository appears to be a starter template created for Task 1 of JP Morgan Chase's software engineering program. It does not contain any specific code or files beyond a basic README. \n\nSome of the key skills that a participant could gain include:\n\n- Version control skills by utilizing GitHub for submission of tasks\n- Configuration management experience setting up a basic repo structure \n- Learning best practices for collaborative coding through use of shared codebase\n\nSince it is simply a template repository without any real implementation, the technical complexity would be considered very low. Participants would likely use it as a starting point to build out and demonstrate their work for", " Here is a summary of the Serhan-Asad/jpmc-task-2 repository:\n\nThis repository appears to be a starter template provided for the second task in JP Morgan Chase's Forage program. The Forage program seems to provide tasks or projects for participants to complete. \n\nThis specific repository does not contain any projects, code, or instructions - it is simply an empty repository intended for participants to add their solution for the second task.\n\nSkills that could be gained by completing the task and adding work to this repository include:\n\n- Programming/coding skills depending on the task requirements \n- Version control skills like committing, pushing code to GitHub\n- Project planning and management \n- Documentation skills if", " Here is a summary of the Serhan-Asad/jpmc-task-3 repository:\n\nThis repository appears to be a starter template for Task 3 of JP Morgan Chase's Forage program. The Forage program seems to be focused on providing learning opportunities in coding/technical skills. \n\nSkills that could be gained by completing the tasks or exercises within this repo include:\n\n- Programming fundamentals (use of variables, functions, conditionals etc depending on the language/framework required)\n\n- Version control with Git/Github for submitting work\n\n- Potentially increased proficiency with a specific language/framework depending on what is required for the tasks \n\n- Practice working through structured coding exercises/problems", ' Here is a summary of the Serhan-Asad/CS1332 repository:\n\nThis repo contains homework assignments from the CS1332 Data Structures and Algorithms course at Georgia Tech. It includes implementations of and practice with:\n\n- HW1: ArrayList - Gained experience implementing a basic array-backed list data structure. Learned about fundamental list operations like add, remove, get. \n\n- HW2: Circular Singly-Linked List - Practiced building a linked list and handling edge cases like wrapping around. Improved skills with pointers and traversal.\n\n- HW3: Stacks & Queues - Built stack and queue abstractions using arrays or linked lists. Learned FIFO', ' Here is a summary of the Serhan-Asad/2340 repository:\n\nThis repository contains a project code named 2340. Unfortunately there is no further information provided in the readme about what the project is or does. \n\nSkills that could potentially be gained from working with this repo include:\n\n- Software development: Building software projects from scratch provides experience with coding, debugging, and problem solving. \n- Version control: Using Git for a personal project introduces skills with tracking changes and collaboration.\n- Project organization: Organizing code and files into a cohesive project structure is a useful skill.\n\nThe technical complexity is difficult to determine without more project details. With no stars and minimal readme information, it is not', ' Here is the summary of the skills gained from this repository:\n\nName: Serhan-Asad/english \nStars: 0\n\nThis repository provides a starting project setup with Create React App. Some of the main skills and concepts learned include:\n\n- Setting up a React project from scratch with Create React App and common NPM scripts. \n\n- Understanding the basic file structure and entry points of a React app.\n\n- Running the project in development and production modes. \n\n- How to test React components.\n\n- Building the app into static files for deployment. \n\n- Options for ejecting from the default CRA setup for more customization.\n\n- Additional resources for learning core React concepts like'] \ No newline at end of file diff --git a/TalentLLM-main/gh_cache/Zaeemahmad34.md b/TalentLLM-main/gh_cache/Zaeemahmad34.md new file mode 100644 index 0000000000000000000000000000000000000000..0637a088a01e8ddab3bf3fa98dbe804cbde1a0dc --- /dev/null +++ b/TalentLLM-main/gh_cache/Zaeemahmad34.md @@ -0,0 +1 @@ +[] \ No newline at end of file diff --git a/TalentLLM-main/gh_cache/danikhan632.md b/TalentLLM-main/gh_cache/danikhan632.md new file mode 100644 index 0000000000000000000000000000000000000000..e49484294e06075b7799a217e877a2da808200bb --- /dev/null +++ b/TalentLLM-main/gh_cache/danikhan632.md @@ -0,0 +1 @@ +[' Here is a summary of the vk-backend-for-triton repository:\n\n- This repository aims to create a Vulkan backend for Triton, enabling Vulkan-compatible devices like those using Apple Metal via MoltenVK to utilize Triton. \n\n- It is currently working on ensuring it produces valid SPIRV assembly code and developing support for Vulkan compute pipelines and Pytorch memory management.\n\n- The skills gained from this repository include learning about backend development for deep learning frameworks, interfacing with GPUs using graphics APIs like Vulkan, and understanding computational graph optimizations.\n\n- Technical complexity is intermediate as it involves developing backend compiler infrastrucutre for a deep learning framework. Understanding of concepts like SPIR-', ' Here is a summary of the iMessage API repository:\n\nThis repository contains a Flask API that allows a user to interact with iMessages on their iPhone/iPad. The key skills gained from working with this repo include:\n\n- API Development with Flask: User would learn how to create RESTful endpoints to perform CRUD operations like sending/retrieving messages and contacts. This introduces Flask framework concepts.\n\n- Working with iPhone Data: User would have to enable full disk access on their Mac and parse iCloud backup data to retrieve messages and contacts. This exposes them to working with iOS application data. \n\n- Cloud backups: User would learn to download, parse and transform iCloud backup files like contacts.vcf into usable formats like', " Here is a summary of the Auto-GPT-AlpacaTrader-Plugin repository:\n\n- The repository contains a plugin that allows an AI assistant created with Anthropic's Constitutional AI technology to integrate with the Alpaca trading platform to perform trades, analyze markets, and manage portfolios.\n\n- Key skills gained include algorithmic trading, financial market data analysis, portfolio management, and integrating an AI model with a trading API. The plugin exposes the Alpaca API to allow placing orders, accessing account data, and retrieving market data.\n\n- The technical complexity is moderate as it requires configuring the Alpaca API key, installing the plugin, and setting up paper trading initially. Basic Python/coding skills would be needed", ' In summary, this Guidance repository provides a framework for structuring and controlling LLMs through templating and programmatic control flow statements. \n\nSome key skills gained:\n\n- Template syntax based on Handlebars to interpolate variables and control program execution\n- Generation control through tags like {{gen}} to produce outputs\n- Logical control flow using tags like {{#if}}, {{#select}}, {{#each}}\n- Structuring multi-step prompts through concepts like hidden blocks, functions, and awaiting missing variables \n- Applying output structure to improve performance on tasks\n- Ensuring valid syntax through techniques like option sets and regex patterns\n- Building role-based conversational models and agent simulations\n- Integration with downstream APIs and tools', ' Here is a summary of the danikhan632/guidance_api repository:\n\n- Technical Complexity: Medium. The repository integrates two text generation tools and abstracts network calls. Some coding/scripting is required. \n\n- Stars: 25 stars. A moderate amount of community interest. \n\n- Purpose: To build an API extension that seamlessly integrates the Guidance large language model into the oobabooga/text-generation-webui interface. \n\n- Key Skills Gained: APIs/extensions development, network requests abstraction, text generation automation, modular programming with Python. \n\n- Benefits: Enables harnessing of advanced LLM capabilities from within a simple interface. Preserves', ' Here is a summary of the Triton programming language repository along with the skills that can be gained from contributing to it:\n\nName: danikhan632/triton \nStars: 0\n\nTriton is an open source programming language and compiler for deep learning workloads targeting GPUs. The repository contains the core compiler infrastructure including the Triton IR dialect, analyses, conversions between dialects, and GPU code generation targets. \n\nSkills gained:\n- Compiler construction: Parsing, IR design, transformations, code generation\n- Domain specific languages: Designing a DSL for deep learning \n- Memory analysis: Analyzing memory access patterns and optimizing data movement\n- GPU programming: Generating code for GPU backends like', " Here is a summary of the danikhan632/Auto-GPT-Messages-Plugin repository:\n\n- The repository contains a plugin that allows Anthropic's Claude AI assistant to send and receive text messages through the iMessage platform using an iMessage API server. \n\n- Skills gained from working with this repo include setting up an iMessage API server, creating a plugin that integrates with an external API, working with messaging platforms and APIs, and customizing Claude's functionality through plugins.\n\n- The technical complexity is moderate as it requires setting up an iMessage server, creating a Python plugin, and integrating with external APIs/platforms. \n\n- 48 stars indicates the plugin is fairly popular and useful for allowing messaging", ' Here is a summary of the Auto-GPT-Text-Gen-Plugin repository:\n\n- Name: Danikhan632/Auto-GPT-Text-Gen-Plugin\n- Stars: 46\n- Purpose: Allows users to fully customize the prompt sent to locally installed large language models (LLMs) for text generation through the Text Generation Web UI (TGWUI). Removes reliance on public models like GPT-3 and GPT-4.\n- How it works: Uses a local TGWUI API to connect to any model running in the TGWUI rather than directly managing models. Provides flexibility to use different models. \n- Skills gained: Customizing text generation prompts, configuring TGWUI'] \ No newline at end of file diff --git a/TalentLLM-main/gh_cache/jryzhik.md b/TalentLLM-main/gh_cache/jryzhik.md new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/TalentLLM-main/gh_cache/nmhossain02.md b/TalentLLM-main/gh_cache/nmhossain02.md new file mode 100644 index 0000000000000000000000000000000000000000..92f4cd06a7ebf3f976fbfda2f900c47bfc8b8c80 --- /dev/null +++ b/TalentLLM-main/gh_cache/nmhossain02.md @@ -0,0 +1 @@ +[' Here is a summary of the BillSplit repository:\n\nThe BillSplit repository is a full-stack web application that allows multiple consumers shopping together to easily split the cost of items in their cart or food order. Through integrating with the NCR Orders API, it allows each consumer to select the items they purchased and directly pay the merchant for their portion of the bill. \n\nKey skills that can be gained from working with this repository include:\n\n- Full-stack web development experience using Flask for the backend and HTML, CSS, JavaScript for the frontend. This provides exposure to both server-side and client-side development.\n\n- Working with retailer APIs like the NCR Orders API to integrate payment processing functionality. This teaches skills around', ' Here is a summary of the nmhossain02/fall-2022-dev-takehome repository:\n\nThis repository contains a coding task to build out a basic todo list application with multiple features. The key skills that can be gained from completing this task include:\n\n- Building forms and accepting user input with features like text fields, tags, and date pickers. \n\n- Rendering lists of data onto the UI and managing state. \n\n- Marking items as complete and filtering/sorting lists.\n\n- Basic CSS styling to display items in cards and differentiate between complete/incomplete.\n\n- Potentially using additional skills like handling arrays, objects, events. \n\nThe technical complexity is on the moderate'] \ No newline at end of file diff --git a/TalentLLM-main/gh_cache/taliu02.md b/TalentLLM-main/gh_cache/taliu02.md new file mode 100644 index 0000000000000000000000000000000000000000..0637a088a01e8ddab3bf3fa98dbe804cbde1a0dc --- /dev/null +++ b/TalentLLM-main/gh_cache/taliu02.md @@ -0,0 +1 @@ +[] \ No newline at end of file diff --git a/TalentLLM-main/gh_cache/tuffstuff9.md b/TalentLLM-main/gh_cache/tuffstuff9.md new file mode 100644 index 0000000000000000000000000000000000000000..91b52d2d14c1996a79a88fa174e7fdfc8560a557 --- /dev/null +++ b/TalentLLM-main/gh_cache/tuffstuff9.md @@ -0,0 +1 @@ +[' Here is a summary of the KeyboardHookLite repository:\n\nThis repository provides a lightweight global keyboard hooking library for .NET applications. It uses low-level PInvoke signatures to hook into keyboard events without relying on outdated .NET classes. \n\nSome key skills and concepts covered include:\n\n- Installing and using a global keyboard hook across applications\n- Understanding low-level unmanaged Windows API calls via PInvoke \n- Processing raw keyboard input events\n- Implementing IDisposable patterns to avoid memory leaks\n- Building extensible event argument classes\n- Supporting modern UI frameworks like WPF, MAUI, WinUI\n\nThe code is well commented and documented to help developers understand how it works under', ' Here is a summary of the NYUCourseTracker repo and the skills that can be gained from it:\n\nThis is a course tracking tool that utilizes Puppeteer to scrape the NYU public course search via a headless browser. It monitors a desired class and sends alerts via Telegram when the class status changes from closed to open, allowing a student to enroll. \n\nThe main skills gained from this project include:\n\n- Web scraping using Puppeteer - Learning how to automate browser actions and extract data programmatically from websites. \n\n- API integration - How to connect apps/scripts to external APIs and services like Telegram. \n\n- Backend development with Node.js - Using Node to build a scraping/monitoring backend.', ' Here is a summary of the Next.js PDF Parser Template repository:\n\nName: nextjs-pdf-parser\nStars: 16\n\nSkills Gained:\n- Working with PDF files in a Next.js application \n- Implementing a file uploader interface using FilePond\n- Parsing PDF content and extracting text using the pdf2json library\n- Resolving technical issues like the "nodeUtil is not defined" error\n- Debugging type definition issues with TypeScript\n\nThis repository provides a template for building a PDF parser application with Next.js. It implements basic PDF uploading and parsing functionality using the FilePond and pdf2json libraries. Working through this template would help develop skills in integrating file uploads', ' Here is a summary of the Next.js 13 with Drizzle ORM & PlanetScale Starter Template repository:\n\nThis repository provides a starter template for building applications with Next.js 13, integrated with the Drizzle ORM for database management and PlanetScale as the database. \n\nSome of the key skills and technologies included in the project are:\n\n- Next.js for server-rendered React apps\n- Drizzle ORM for efficient database operations \n- PlanetScale for a scalable serverless database\n- TailwindCSS for rapid UI development\n- Server Actions for improved server-side functionality\n- An experimental hook for form state management\n\nThe technical complexity is moderate as it includes setting up backend integrations like the database.', ' Here is the summary of the react-blackjack repository:\n\nSkills gained: React, Node.js, Express, MongoDB, Passport authentication, deploying to Heroku. \n\nThis project allows users to play blackjack against the dealer. It uses React for the front-end UI and interactions. Node.js and Express are used to build the backend API. MongoDB stores user accounts and balances so data is persistent. Passport handles authentication so users can create accounts. Code is deployed to Heroku so the app can be run remotely.\n\nThe tech stack incorporates modern JavaScript tools like React, Node, and MongoDB. Building the full-stack app from front-end to backend provides exposure to developing both sides. Implementing', ' Here is a summary of the Tailwind Input Tag Component repository:\n\nThis is a React component that allows users to easily create tags using Tailwind CSS for styling. It provides a lightweight tag input that users can add tags to by typing and pressing enter or tab. Tags can be removed by backspacing on an empty input. \n\nKey features include being lightweight, styled with Tailwind for consistency, adding/removing tags via keyboard shortcuts, and option to set a maximum number of tags. \n\nThe component file is designed to be imported and used within other projects. Usage involves importing the component, setting initial state, and passing callbacks to handle tag changes. \n\nProps allow customizing the input name, placeholder, initial value,', ' Here is a summary of the visualizer.coffee Shot Downloader repository:\n\nName: tuffstuff9/visualizer.coffee-shot-downloader\nStars: 1\n\nThis is a simple Python Jupyter notebook script that allows downloading of all shot data from visualizer.coffee as JSON files. Some key points:\n\n- Skills Gained: API interfacing, asynchronous HTTP requests, data scraping, JSON parsing\n- Technical Complexity: Beginner - uses basic Python and common libraries like requests and aiohttp \n- Functionality: Logs in using credentials, fetches all shot IDs, asynchronously downloads shot data as JSON files\n- Limitations: Only downloads data, no actual shot files. No filtering of shots.', " Here is a summary of the tuffstuff9/Copy-Button-for-ChatGPT repository:\n\nThe repository adds a native looking copy button to ChatGPT messages. This allows users to easily copy the text from ChatGPT's responses. \n\nSkills Gained:\n- Front-end web development by adding UI elements \n- Browser extension development using the ChatGPT extension API\n- Basic JavaScript and HTML/CSS skills\n\nTechnical Complexity: Beginner. Includes simple JavaScript and HTML/CSS code to add a button. \n\nPotential Benefits for Users: Allows quicker copying of ChatGPT responses without having to manually select text. Provides a more seamless user experience within the ChatGPT"] \ No newline at end of file diff --git a/TalentLLM-main/main.py b/TalentLLM-main/main.py new file mode 100644 index 0000000000000000000000000000000000000000..1b514ff8f1250d10d2555e246440bbc37eb43e13 --- /dev/null +++ b/TalentLLM-main/main.py @@ -0,0 +1,50 @@ +import gspread +import os,random +from Candidate import JobCandidate +from dotenv import load_dotenv +from compator import bubble_sort +from results import writeToSheets +from resume_conversation import chat_with_candidate +sa = gspread.service_account(filename='service_creds.json') +sh = sa.open("Figma_swe") +load_dotenv() + +wks = sh.worksheet("Sheet1") +data = wks.get_all_values() + + +# Load environment variables from the .env file +load_dotenv() +# destination_path = os.path.join(os.getcwd(), id) + +candidates=[] + +# os.environ['COMPARATOR_LLM']="chat-bison" +os.environ['COMPARATOR_LLM']="gpt-3.5-turbo-1106" +for i in range(1, 7): + candid =JobCandidate(data[i]) + candidates.append(candid) + +# random.shuffle(candidates) +sort_cand = bubble_sort(candidates) + +writeToSheets(candidates) + +for idx, candidate in enumerate(sort_cand): + print(str(idx) + '. ' + candidate.email) + +print('Select a candidate to chat with. Type in their index number. Type -1 if you dont want to chat.') +idx = int(input()) +if idx != -1: + selected_candidate = candidates[idx] + chat_with_candidate(selected_candidate) + +# for candidate in candidates: +# print(candidate) +# print() # Print a blank line between candidates for better readability + + + + + + diff --git a/TalentLLM-main/mathpix.py b/TalentLLM-main/mathpix.py new file mode 100644 index 0000000000000000000000000000000000000000..98ba335a7f8c8232c519b27dbdb3177f74e0b6ed --- /dev/null +++ b/TalentLLM-main/mathpix.py @@ -0,0 +1,62 @@ + + +import requests +import time +import json +import os +from dotenv import load_dotenv +load_dotenv() + +HEADERS = { + 'app_id': os.environ.get('MATHPIX_APP_ID', 'default_app_id'), + 'app_key': os.environ.get('MATHPIX_APP_KEY', 'default_app_key') +} + + + +def extract_text(file_path: str) -> str: + print("Parsing resume") + if not os.path.exists(file_path): + raise FileNotFoundError(f"The file at {file_path} does not exist.") + + file_name = os.path.basename(file_path) + + url1 = 'https://api.mathpix.com/v3/pdf' + + with open(file_path, 'rb') as file: + files = {'file': file} + data = {'options_json': json.dumps({ + "conversion_formats": {"md": True}, + "math_inline_delimiters": ["$", "$"], + "rm_spaces": True + })} + status_resp = requests.post(url1, headers=HEADERS, files=files, data=data) + + if status_resp.status_code != 200: + raise Exception(f"Failed to upload PDF: {status_resp.text}") + + status_resp_data = status_resp.json() + pdf_id = status_resp_data.get('pdf_id') + + if not pdf_id: + raise Exception("Failed to retrieve PDF ID from response.") + + time.sleep(1) + + url2 = f'https://api.mathpix.com/v3/pdf/{pdf_id}' + while True: + challenge_resp = requests.get(url2, headers=HEADERS) + challenge_resp_data = challenge_resp.json() + if challenge_resp_data.get('status') == 'completed': + break + time.sleep(1) + + url3 = f'https://api.mathpix.com/v3/pdf/{pdf_id}.mmd' + contents = requests.get(url3, headers=HEADERS) + + if contents.status_code != 200: + raise Exception(f"Failed to download converted file: {contents.text}") + + open(os.path.join(os.getcwd(),"resume_mmds", (str(file_name)+'.mmd')),"w").write(contents.text) + + return contents.text diff --git a/TalentLLM-main/notebooks/cheap_compartor.py b/TalentLLM-main/notebooks/cheap_compartor.py new file mode 100644 index 0000000000000000000000000000000000000000..dc9251e63ee92d71acf06f81be5889d033f73b1e --- /dev/null +++ b/TalentLLM-main/notebooks/cheap_compartor.py @@ -0,0 +1,9 @@ +from litellm import embedding +import os + + +# print(help(embedding)) + +response = embedding('huggingface/microsoft/codebert-base', input=["good morning from litellm"]) +# print(len(response['data'][0]["embeddingtext-embedding-ada-002"])) +print(int(response['usage']["total_tokens"])) \ No newline at end of file diff --git a/TalentLLM-main/notebooks/comparisons copy 2.json b/TalentLLM-main/notebooks/comparisons copy 2.json new file mode 100644 index 0000000000000000000000000000000000000000..61b750a085287455579c9460211d2aa1cd6124c3 --- /dev/null +++ b/TalentLLM-main/notebooks/comparisons copy 2.json @@ -0,0 +1,130 @@ +{ + "danikhan632@gmail.com#saad.mufti@mit.edu#0":-1, + "saad.mufti@mit.edu#danikhan632@gmail.com#0":1, + "saad.mufti@mit.edu#mr.zeyadalsakhi@gmail.com#0":-1, + "mr.zeyadalsakhi@gmail.com#saad.mufti@mit.edu#0":1, + "mr.zeyadalsakhi@gmail.com#azishaqui@gmail.com#0":-1, + "azishaqui@gmail.com#mr.zeyadalsakhi@gmail.com#0":1, + "azishaqui@gmail.com#dan.hus120@gmail.com#0":-1, + "dan.hus120@gmail.com#azishaqui@gmail.com#0":1, + "dan.hus120@gmail.com#tawfiqaliu@gmail.com#0":1, + "tawfiqaliu@gmail.com#dan.hus120@gmail.com#0":-1, + "dan.hus120@gmail.com#nhossain34@gatech.edu#0":1, + "nhossain34@gatech.edu#dan.hus120@gmail.com#0":-1, + "dan.hus120@gmail.com#pyudhistira02@gmail.com#0":-1, + "pyudhistira02@gmail.com#dan.hus120@gmail.com#0":1, + "pyudhistira02@gmail.com#sasad3@gatech.edu#0":-1, + "sasad3@gatech.edu#pyudhistira02@gmail.com#0":1, + "sasad3@gatech.edu#smufti3@gatech.edu#0":1, + "smufti3@gatech.edu#sasad3@gatech.edu#0":-1, + "sasad3@gatech.edu#jamesryzhkov@gmail.com#0":1, + "jamesryzhkov@gmail.com#sasad3@gatech.edu#0":-1, + "sasad3@gatech.edu#tawfiqfm@gmail.com#0":1, + "tawfiqfm@gmail.com#sasad3@gatech.edu#0":-1, + "sasad3@gatech.edu#Zahmad34@gatech.edu#0":1, + "Zahmad34@gatech.edu#sasad3@gatech.edu#0":-1, + "azishaqui@gmail.com#tawfiqaliu@gmail.com#0":-1, + "tawfiqaliu@gmail.com#azishaqui@gmail.com#0":1, + "tawfiqaliu@gmail.com#nhossain34@gatech.edu#0":-1, + "nhossain34@gatech.edu#tawfiqaliu@gmail.com#0":1, + "pyudhistira02@gmail.com#smufti3@gatech.edu#0":1, + "smufti3@gatech.edu#pyudhistira02@gmail.com#0":-1, + "pyudhistira02@gmail.com#jamesryzhkov@gmail.com#0":-1, + "jamesryzhkov@gmail.com#pyudhistira02@gmail.com#0":1, + "jamesryzhkov@gmail.com#tawfiqfm@gmail.com#0":1, + "tawfiqfm@gmail.com#jamesryzhkov@gmail.com#0":-1, + "jamesryzhkov@gmail.com#Zahmad34@gatech.edu#0":-1, + "Zahmad34@gatech.edu#jamesryzhkov@gmail.com#0":1, + "dan.hus120@gmail.com#smufti3@gatech.edu#0":1, + "smufti3@gatech.edu#dan.hus120@gmail.com#0":-1, + "pyudhistira02@gmail.com#tawfiqfm@gmail.com#0":1, + "tawfiqfm@gmail.com#pyudhistira02@gmail.com#0":-1, + "nhossain34@gatech.edu#smufti3@gatech.edu#0":-1, + "smufti3@gatech.edu#nhossain34@gatech.edu#0":1, + "dan.hus120@gmail.com#tawfiqfm@gmail.com#0":1, + "tawfiqfm@gmail.com#dan.hus120@gmail.com#0":1, + "dan.hus120@gmail.com#mr.zeyadalsakhi@gmail.com#0":-1, + "mr.zeyadalsakhi@gmail.com#dan.hus120@gmail.com#0":1, + "mr.zeyadalsakhi@gmail.com#danikhan632@gmail.com#0":1, + "danikhan632@gmail.com#mr.zeyadalsakhi@gmail.com#0":1, + "danikhan632@gmail.com#tawfiqfm@gmail.com#0":1, + "tawfiqfm@gmail.com#danikhan632@gmail.com#0":1, + "sasad3@gatech.edu#azishaqui@gmail.com#0":-1, + "azishaqui@gmail.com#sasad3@gatech.edu#0":1, + "pyudhistira02@gmail.com#Zahmad34@gatech.edu#0":1, + "Zahmad34@gatech.edu#pyudhistira02@gmail.com#0":-1, + "saad.mufti@mit.edu#sasad3@gatech.edu#0":1, + "sasad3@gatech.edu#saad.mufti@mit.edu#0":-1, + "saad.mufti@mit.edu#azishaqui@gmail.com#0":-1, + "azishaqui@gmail.com#saad.mufti@mit.edu#0":1, + "azishaqui@gmail.com#jamesryzhkov@gmail.com#0":-1, + "jamesryzhkov@gmail.com#azishaqui@gmail.com#0":1, + "jamesryzhkov@gmail.com#smufti3@gatech.edu#0":-1, + "smufti3@gatech.edu#jamesryzhkov@gmail.com#0":1, + "pyudhistira02@gmail.com#nhossain34@gatech.edu#0":-1, + "nhossain34@gatech.edu#pyudhistira02@gmail.com#0":1, + "dan.hus120@gmail.com#Zahmad34@gatech.edu#0":-1, + "Zahmad34@gatech.edu#dan.hus120@gmail.com#0":1, + "mr.zeyadalsakhi@gmail.com#jamesryzhkov@gmail.com#0":-1, + "jamesryzhkov@gmail.com#mr.zeyadalsakhi@gmail.com#0":1, + "Zahmad34@gatech.edu#nhossain34@gatech.edu#0":-1, + "nhossain34@gatech.edu#Zahmad34@gatech.edu#0":1, + "nhossain34@gatech.edu#danikhan632@gmail.com#0":1, + "danikhan632@gmail.com#nhossain34@gatech.edu#0":1, + "danikhan632@gmail.com#sasad3@gatech.edu#0":-1, + "sasad3@gatech.edu#danikhan632@gmail.com#0":1, + "mr.zeyadalsakhi@gmail.com#pyudhistira02@gmail.com#0":1, + "pyudhistira02@gmail.com#mr.zeyadalsakhi@gmail.com#0":-1, + "danikhan632@gmail.com#smufti3@gatech.edu#0":1, + "smufti3@gatech.edu#danikhan632@gmail.com#0":1, + "Zahmad34@gatech.edu#azishaqui@gmail.com#0":-1, + "azishaqui@gmail.com#Zahmad34@gatech.edu#0":1, + "azishaqui@gmail.com#nhossain34@gatech.edu#0":-1, + "nhossain34@gatech.edu#azishaqui@gmail.com#0":1, + "nhossain34@gatech.edu#sasad3@gatech.edu#0":-1, + "sasad3@gatech.edu#nhossain34@gatech.edu#0":1, + "sasad3@gatech.edu#mr.zeyadalsakhi@gmail.com#0":-1, + "mr.zeyadalsakhi@gmail.com#sasad3@gatech.edu#0":1, + "jamesryzhkov@gmail.com#tawfiqaliu@gmail.com#0":-1, + "tawfiqaliu@gmail.com#jamesryzhkov@gmail.com#0":1, + "dan.hus120@gmail.com#saad.mufti@mit.edu#0":0, + "saad.mufti@mit.edu#dan.hus120@gmail.com#0":0, + "saad.mufti@mit.edu#smufti3@gatech.edu#0":-1, + "smufti3@gatech.edu#saad.mufti@mit.edu#0":1, + "smufti3@gatech.edu#tawfiqfm@gmail.com#0":-1, + "tawfiqfm@gmail.com#smufti3@gatech.edu#0":1, + "mr.zeyadalsakhi@gmail.com#tawfiqaliu@gmail.com#0":1, + "tawfiqaliu@gmail.com#mr.zeyadalsakhi@gmail.com#0":-1, + "mr.zeyadalsakhi@gmail.com#Zahmad34@gatech.edu#0":1, + "Zahmad34@gatech.edu#mr.zeyadalsakhi@gmail.com#0":-1, + "azishaqui@gmail.com#danikhan632@gmail.com#0":1, + "danikhan632@gmail.com#azishaqui@gmail.com#0":1, + "saad.mufti@mit.edu#Zahmad34@gatech.edu#0":-1, + "Zahmad34@gatech.edu#saad.mufti@mit.edu#0":1, + "smufti3@gatech.edu#tawfiqaliu@gmail.com#0":0, + "tawfiqaliu@gmail.com#smufti3@gatech.edu#0":0, + "smufti3@gatech.edu#mr.zeyadalsakhi@gmail.com#0":-1, + "mr.zeyadalsakhi@gmail.com#smufti3@gatech.edu#0":1, + "mr.zeyadalsakhi@gmail.com#tawfiqfm@gmail.com#0":-1, + "tawfiqfm@gmail.com#mr.zeyadalsakhi@gmail.com#0":1, + "tawfiqfm@gmail.com#tawfiqaliu@gmail.com#0":-1, + "tawfiqaliu@gmail.com#tawfiqfm@gmail.com#0":1, + "dan.hus120@gmail.com#danikhan632@gmail.com#0":1, + "danikhan632@gmail.com#dan.hus120@gmail.com#0":-1, + "Zahmad34@gatech.edu#tawfiqaliu@gmail.com#0":-1, + "tawfiqaliu@gmail.com#Zahmad34@gatech.edu#0":1, + "tawfiqaliu@gmail.com#danikhan632@gmail.com#0":1, + "danikhan632@gmail.com#tawfiqaliu@gmail.com#0":1, + "danikhan632@gmail.com#pyudhistira02@gmail.com#0":-1, + "pyudhistira02@gmail.com#danikhan632@gmail.com#0":1, + "nhossain34@gatech.edu#mr.zeyadalsakhi@gmail.com#0":-1, + "mr.zeyadalsakhi@gmail.com#nhossain34@gatech.edu#0":1, + "jamesryzhkov@gmail.com#nhossain34@gatech.edu#0":-1, + "nhossain34@gatech.edu#jamesryzhkov@gmail.com#0":1, + "nhossain34@gatech.edu#tawfiqfm@gmail.com#0":-1, + "tawfiqfm@gmail.com#nhossain34@gatech.edu#0":1, + "pyudhistira02@gmail.com#tawfiqaliu@gmail.com#0":1, + "tawfiqaliu@gmail.com#pyudhistira02@gmail.com#0":-1, + "smufti3@gatech.edu#Zahmad34@gatech.edu#0":1, + "Zahmad34@gatech.edu#smufti3@gatech.edu#0":-1 + } \ No newline at end of file diff --git a/TalentLLM-main/notebooks/comparisons copy 3.json b/TalentLLM-main/notebooks/comparisons copy 3.json new file mode 100644 index 0000000000000000000000000000000000000000..22cfd6b3bbbd970e756ec489104f875d718d2458 --- /dev/null +++ b/TalentLLM-main/notebooks/comparisons copy 3.json @@ -0,0 +1 @@ +{"smufti3@gatech.edu#tawfiqaliu@gmail.com#0": 1, "tawfiqaliu@gmail.com#smufti3@gatech.edu#0": -1, "smufti3@gatech.edu#tawfiqfm@gmail.com#0": 1, "tawfiqfm@gmail.com#smufti3@gatech.edu#0": -1, "smufti3@gatech.edu#danikhan632@gmail.com#0": 1, "danikhan632@gmail.com#smufti3@gatech.edu#0": -1, "smufti3@gatech.edu#jamesryzhkov@gmail.com#0": 1, "jamesryzhkov@gmail.com#smufti3@gatech.edu#0": -1, "tawfiqaliu@gmail.com#tawfiqfm@gmail.com#0": 1, "tawfiqfm@gmail.com#tawfiqaliu@gmail.com#0": -1, "tawfiqaliu@gmail.com#danikhan632@gmail.com#0": 1, "danikhan632@gmail.com#tawfiqaliu@gmail.com#0": -1, "tawfiqaliu@gmail.com#jamesryzhkov@gmail.com#0": 1, "jamesryzhkov@gmail.com#tawfiqaliu@gmail.com#0": -1, "tawfiqfm@gmail.com#danikhan632@gmail.com#0": 1, "danikhan632@gmail.com#tawfiqfm@gmail.com#0": -1, "tawfiqfm@gmail.com#jamesryzhkov@gmail.com#0": 1, "jamesryzhkov@gmail.com#tawfiqfm@gmail.com#0": -1, "danikhan632@gmail.com#jamesryzhkov@gmail.com#0": 1, "jamesryzhkov@gmail.com#danikhan632@gmail.com#0": -1} \ No newline at end of file diff --git a/TalentLLM-main/notebooks/compator.py b/TalentLLM-main/notebooks/compator.py new file mode 100644 index 0000000000000000000000000000000000000000..199d823291c8c15c86cb717e6f7f872252331775 --- /dev/null +++ b/TalentLLM-main/notebooks/compator.py @@ -0,0 +1,132 @@ +import openai; +import json, os,sys +from dotenv import load_dotenv +load_dotenv() +openai.api_key = os.environ.get("OPENAI_API_KEY") +from Candidate import JobCandidate +def printc(obj, color="cyan"): + color_code = { + "black": "30", "red": "31", "green": "32", "yellow": "33", + "blue": "34", "magenta": "35", "cyan": "36", "white": "37" + } + colored_text = f"\033[{color_code[color]}m{obj}\033[0m" if color in color_code else obj + print(colored_text) + + +LLM=os.environ.get("COMPARATOR_LLM","gpt-4-0613") +# LLM=os.environ.get("COMPARATOR_LLM","gpt-3.5-turbo-1106") +def getContent(resumeA: str, resumeB: str) -> str: + return ( + "Given the following two SWE candidates, choose between the two. Here is the rubric: " + + get_rubric() + + "Candidate A: " + + "\nRESUME:\n" +resumeA+"\nEND Resume\n" + + " END OF Candidate A" + + "\n\nCandidate B: " + + "\nRESUME:\n" +resumeB+"\nEND Resume\n" + + " END OF Candidate B" + ) + + + +def compare_resumes(content:str, nameA="", nameB=""): + choice =0 + response = openai.ChatCompletion.create( + model=LLM, + messages=[{"role": "user", "content": content}], + + functions=[ + { + "name": "selectCanidate", + "description": "choose between the two canidates", + "parameters": { + "type": "object", + "properties": { + "choice_num": { + "type": "integer", + "description": "1 for Candidate A is the best fit, 2 for Candidate B is the best fit", + "required": ["choice_num"], + }, + "justifcation": { + "type": "string", + "description": "justifcation for why you chose the candidate max 25 words", + "required": ["justifcation"], + }, + } + }, + } + ], + function_call="auto", + + + ) + + message = response["choices"][0]["message"] + + if message.get("function_call"): + function_name = message["function_call"]["name"] + try: + function_args = json.loads(message["function_call"]["arguments"]) + choice = (int(function_args["choice_num"])) + except: + printc("eroor","red") + printc(message["function_call"],'red') + + return 1 + if function_name == "selectCanidate": + + if choice==1: + printc(nameA+" wins over "+nameB,"cyan") + elif choice==2: + printc(nameB+" wins over "+nameA,"green") + + printc(function_args["justifcation"],"yellow") + + return choice + + +def get_rubric(): + text = open("rubric.txt","r").read() + return "\nRubric:\n" +str(text)+"\nEND Rubric\n" + + + + + +def comp(candidateA:JobCandidate, candidateB:JobCandidate, rub_id:int=0 ) -> int: + comp_table= json.load(open("comparisons.json","r")) + tag= (candidateA.email+"#"+candidateB.email+"#"+str(rub_id)) + inv_tag= (candidateB.email+"#"+candidateA.email+"#"+str(rub_id)) + if tag in comp_table: + return comp_table[tag] + elif inv_tag in comp_table: + return comp_table[inv_tag] * -1 + else: + choice = compare_resumes(getContent(candidateA.resume_text, candidateB.resume_text), candidateA.name, candidateB.name) + if choice == 1: + choice = -1 + elif choice == 2: + choice = 1 + comp_table[tag]=choice + + json.dump(comp_table, open("comparisons.json","w")) + + return choice + + +def bubble_sort(candidates: list) -> list: + n = len(candidates) + for i in range(n): + swapped = False + for j in range(0, n-i-1): + if candidates[j].email == candidates[j+1].email: + continue + elif comp(candidates[j], candidates[j+1]) > 0: + candidates[j], candidates[j+1] = candidates[j+1], candidates[j] + swapped = True + if not swapped: + break + + + return candidates + diff --git a/TalentLLM-main/notebooks/compator_parallel.py b/TalentLLM-main/notebooks/compator_parallel.py new file mode 100644 index 0000000000000000000000000000000000000000..0ce9c7c61356022255d4f3db3896ada4148eae95 --- /dev/null +++ b/TalentLLM-main/notebooks/compator_parallel.py @@ -0,0 +1,110 @@ +import openai; +import json, os, threading + +from dotenv import load_dotenv +load_dotenv() +openai.api_key = os.environ.get("OPENAI_API_KEY") +from Candidate import JobCandidate + +def getContent(resumeA: str, resumeB: str) -> str: + return ( + "Given the following two SWE candidates, choose between the two. Here is the rubric: " + + get_rubric() + + "Candidate A: " + + "\nRESUME:\n" +resumeA+"\nEND Resume\n" + + " END OF Candidate A" + + "\n\nCandidate B: " + + "\nRESUME:\n" +resumeB+"\nEND Resume\n" + + " END OF Candidate B" + ) + + + +def compare_resumes(content:str): + choice =0 + response = openai.ChatCompletion.create( + model="gpt-4-0613", + messages=[{"role": "user", "content": content}], + functions=[ + { + "name": "selectCanidate", + "description": "choose between the two canidates", + "parameters": { + "type": "object", + "properties": { + "choice_num": { + "type": "integer", + "description": "1 for Candidate A is the best fit, 2 for Candidate B is the best fit", + "required": ["choice_num"], + }, + "justifcation": { + "type": "string", + "description": "justifcation for why you chose the candidate", + "required": ["justifcation"], + }, + } + }, + } + ], + function_call="auto", + + ) + + message = response["choices"][0]["message"] + + if message.get("function_call"): + function_name = message["function_call"]["name"] + function_args = json.loads(message["function_call"]["arguments"]) + if function_name == "selectCanidate": + choice = (int(function_args["choice_num"])) + print(function_args["justifcation"]) + + return choice + + +def get_rubric(): + text = open("rubric.txt","r").read() + return "\nRubric:\n" +str(text)+"\nEND Rubric\n" + + + + +def comp_parallel(candidateA: JobCandidate, candidateB: JobCandidate, rub_id: int, comp_table: dict, lock: threading.Lock): + tag = f"{candidateA.email}#{candidateB.email}#{rub_id}" + if tag not in comp_table: + choice = compare_resumes(getContent(candidateA.resume_text, candidateB.resume_text)) + if choice == 1: + choice = -1 + elif choice == 2: + choice = 1 + + with lock: + comp_table[tag] = choice + + +def pre_compute_comparisons(candidates: list, rub_id: int = 0) -> dict: + comp_table= json.load(open("comparisons.json","r")) + lock = threading.Lock() + threads = [] + + for i in range(len(candidates)): + for j in range(i + 1, len(candidates)): + thread = threading.Thread(target=comp_parallel, args=(candidates[i], candidates[j], rub_id, comp_table, lock)) + threads.append(thread) + thread.start() + + for thread in threads: + thread.join() + + json.dump(comp_table, open("comparisons.json","w")) + return comp_table + +def bubble_sort(candidates: list, rub_id: int = 0) -> list: + n = len(candidates) + comp_table = pre_compute_comparisons(candidates, rub_id) + for i in range(n): + for j in range(n - i - 1): + tag = f"{candidates[j].email}#{candidates[j + 1].email}#"+str(rub_id) + if comp_table[tag] > 0: + candidates[j], candidates[j + 1] = candidates[j + 1], candidates[j] + return candidates diff --git a/TalentLLM-main/notebooks/extractor.py b/TalentLLM-main/notebooks/extractor.py new file mode 100644 index 0000000000000000000000000000000000000000..7f52c3df74f0e9745b4054c84b801e2e468c0497 --- /dev/null +++ b/TalentLLM-main/notebooks/extractor.py @@ -0,0 +1,45 @@ +text = open("resume_mmds/11NP1Fgf1hOP6pmX0HTbnq7GWY6eAyw4Y.pdf.mmd","r").read() +import json, os +import openai +from dotenv import load_dotenv +load_dotenv() +# openai.api_key = os.environ.get("OPENAI_API_KEY") +openai.api_base = "http://0.0.0.0:8080" + +def extract(content:str): + choice =0 + response = openai.ChatCompletion.create( + model="test", + messages=[{"role": "user", "content": content}], + functions=[ + { + "name": "infoExtract", + "description": "extract info from resume", + "parameters": { + "type": "object", + "properties": { + "linkedin_url": {"type": "string"}, + "portfolio_url": {"type": "string"}, + "github_url": {"type": "string"}, + "stackoverflow_url": {"type": "string"}, + "name": {"type": "string"}, + + } + }, + } + ], + function_call="auto", + + ) + + message = response["choices"][0]["message"] + + if message.get("function_call"): + function_name = message["function_call"]["name"] + function_args = json.loads(message["function_call"]["arguments"]) + if function_name == "infoExtract": + print(function_args) + + + +extract(text) diff --git a/TalentLLM-main/notebooks/hardset.ipynb b/TalentLLM-main/notebooks/hardset.ipynb new file mode 100644 index 0000000000000000000000000000000000000000..0c64fcbf86bb692c0a5b33d19fa7e6bd9bceda42 --- /dev/null +++ b/TalentLLM-main/notebooks/hardset.ipynb @@ -0,0 +1,21 @@ +{ + "cells": [ + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "\n" + ] + } + ], + "metadata": { + "language_info": { + "name": "python" + }, + "orig_nbformat": 4 + }, + "nbformat": 4, + "nbformat_minor": 2 +} diff --git a/TalentLLM-main/notebooks/rag.ipynb b/TalentLLM-main/notebooks/rag.ipynb new file mode 100644 index 0000000000000000000000000000000000000000..8e683baef4b47b48a915c50b4327bef56298112c --- /dev/null +++ b/TalentLLM-main/notebooks/rag.ipynb @@ -0,0 +1,449 @@ +{ + "cells": [ + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "# Redis LangChain OpenAI eCommerce Chatbot" + ] + }, + { + "cell_type": "code", + "execution_count": 1, + "metadata": { + "colab": { + "base_uri": "https://localhost:8080/" + }, + "id": "5-h_nDGp3Kdf", + "outputId": "94191443-3844-4c1d-a26f-7619d976a55b", + "tags": [] + }, + "outputs": [ + { + "name": "stdout", + "output_type": "stream", + "text": [ + "/usr/bin/zsh: /home/green/miniconda3/lib/libtinfo.so.6: no version information available (required by /usr/bin/zsh)\n", + "\n", + "\u001b[1m[\u001b[0m\u001b[34;49mnotice\u001b[0m\u001b[1;39;49m]\u001b[0m\u001b[39;49m A new release of pip is available: \u001b[0m\u001b[31;49m23.1.2\u001b[0m\u001b[39;49m -> \u001b[0m\u001b[32;49m23.3.1\u001b[0m\n", + "\u001b[1m[\u001b[0m\u001b[34;49mnotice\u001b[0m\u001b[1;39;49m]\u001b[0m\u001b[39;49m To update, run: \u001b[0m\u001b[32;49mpip install --upgrade pip\u001b[0m\n" + ] + } + ], + "source": [ + "# Install requirements\n", + "!pip install -r requirements.txt" + ] + }, + { + "cell_type": "code", + "execution_count": 2, + "metadata": { + "tags": [] + }, + "outputs": [ + { + "name": "stdout", + "output_type": "stream", + "text": [ + "/usr/bin/zsh: /home/green/miniconda3/lib/libtinfo.so.6: no version information available (required by /usr/bin/zsh)\n", + "/home/linuxbrew/.linuxbrew/opt/python@3.11/lib/python3.11/site-packages/gdown/cli.py:126: FutureWarning: Option `--id` was deprecated in version 4.3.1 and will be removed in 5.0. You don't need to pass it anymore to use a file ID.\n", + " warnings.warn(\n", + "Downloading...\n", + "From (uriginal): https://drive.google.com/uc?id=1tHWB6u3yQCuAgOYc-DxtZ8Mru3uV5_lj\n", + "From (redirected): https://drive.google.com/uc?id=1tHWB6u3yQCuAgOYc-DxtZ8Mru3uV5_lj&confirm=t&uuid=f678b48d-4f3e-44f9-bf60-03ca828cb67c\n", + "To: /home/green/code/gatech/ai_atl/inital_work/product_data.csv\n", + "100%|████████████████████████████████████████| 225M/225M [00:09<00:00, 24.0MB/s]\n" + ] + } + ], + "source": [ + "# Download the dataset\n", + "!gdown --id 1tHWB6u3yQCuAgOYc-DxtZ8Mru3uV5_lj" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "## Preprocess dataset" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": { + "tags": [] + }, + "outputs": [], + "source": [ + "import pandas as pd\n", + "\n", + "MAX_TEXT_LENGTH=512\n", + "\n", + "def auto_truncate(val):\n", + " \"\"\"Truncate the given text.\"\"\"\n", + " return val[:MAX_TEXT_LENGTH]\n", + "\n", + "# Load Product data and truncate long text fields\n", + "all_prods_df = pd.read_csv(\"product_data.csv\", converters={\n", + " 'bullet_point': auto_truncate,\n", + " 'item_keywords': auto_truncate,\n", + " 'item_name': auto_truncate\n", + "})" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": { + "colab": { + "base_uri": "https://localhost:8080/", + "height": 669 + }, + "id": "00_n4VWH7FoB", + "outputId": "f26daa8c-4af9-4def-d5ab-3197777fe2f9", + "tags": [] + }, + "outputs": [], + "source": [ + "# Contruct a primary key from item ID and domain name\n", + "all_prods_df['primary_key'] = (\n", + " all_prods_df['item_id'] + '-' + all_prods_df['domain_name']\n", + ")\n", + "# Replace empty strings with None and drop\n", + "all_prods_df['item_keywords'].replace('', None, inplace=True)\n", + "all_prods_df.dropna(subset=['item_keywords'], inplace=True)\n", + "\n", + "# Reset pandas dataframe index\n", + "all_prods_df.reset_index(drop=True, inplace=True)\n", + "\n", + "all_prods_df.head()" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": { + "tags": [] + }, + "outputs": [], + "source": [ + "# Num products to use (subset)\n", + "NUMBER_PRODUCTS = 2500 \n", + "\n", + "# Get the first 1000 products with non-empty item keywords\n", + "product_metadata = ( \n", + " all_prods_df\n", + " .head(NUMBER_PRODUCTS)\n", + " .to_dict(orient='index')\n", + ")" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": { + "id": "Iw7rlppY8f3a", + "tags": [] + }, + "outputs": [], + "source": [ + "# Check one of the products\n", + "product_metadata[0]" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "## Set up Redis as a vector db" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": { + "tags": [] + }, + "outputs": [], + "source": [ + "from langchain.embeddings import OpenAIEmbeddings\n", + "from langchain.vectorstores.redis import Redis as RedisVectorStore\n", + "\n", + "# data that will be embedded and converted to vectors\n", + "texts = [\n", + " v['item_name'] for k, v in product_metadata.items()\n", + "]\n", + "\n", + "# product metadata that we'll store along our vectors\n", + "metadatas = list(product_metadata.values())\n", + "\n", + "# we will use OpenAI as our embeddings provider\n", + "embedding = OpenAIEmbeddings()\n", + "\n", + "# name of the Redis search index to create\n", + "index_name = \"products\"\n", + "\n", + "# assumes you have a redis stack server running on within your docker compose network\n", + "redis_url = \"redis://redis:6379\"\n", + "\n", + "# create and load redis with documents\n", + "vectorstore = RedisVectorStore.from_texts(\n", + " texts=texts,\n", + " metadatas=metadatas,\n", + " embedding=embedding,\n", + " index_name=index_name,\n", + " redis_url=redis_url\n", + ")" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "## Build the ChatBot with ConversationalRetrieverChain" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": { + "tags": [] + }, + "outputs": [], + "source": [ + "from langchain.callbacks.base import CallbackManager\n", + "from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler\n", + "from langchain.chains import (\n", + " ConversationalRetrievalChain,\n", + " LLMChain\n", + ")\n", + "from langchain.chains.question_answering import load_qa_chain\n", + "from langchain.llms import OpenAI\n", + "from langchain.prompts.prompt import PromptTemplate\n", + "\n", + "template = \"\"\"Given the following chat history and a follow up question, rephrase the follow up input question to be a standalone question.\n", + "Or end the conversation if it seems like it's done.\n", + "\n", + "Chat History:\\\"\"\"\n", + "{chat_history}\n", + "\\\"\"\"\n", + "\n", + "Follow Up Input: \\\"\"\"\n", + "{question}\n", + "\\\"\"\"\n", + "\n", + "Standalone question:\"\"\"\n", + "\n", + "condense_question_prompt = PromptTemplate.from_template(template)\n", + "\n", + "template = \"\"\"You are a friendly, conversational retail shopping assistant. Use the following context including product names, descriptions, and keywords to show the shopper whats available, help find what they want, and answer any questions.\n", + "It's ok if you don't know the answer.\n", + "\n", + "Context:\\\"\"\"\n", + "{context}\n", + "\\\"\"\"\n", + "\n", + "Question:\\\"\n", + "\\\"\"\"\n", + "\n", + "Helpful Answer:\"\"\"\n", + "\n", + "qa_prompt= PromptTemplate.from_template(template)\n", + "\n", + "\n", + "# define two LLM models from OpenAI\n", + "llm = OpenAI(temperature=0)\n", + "\n", + "streaming_llm = OpenAI(\n", + " streaming=True,\n", + " callback_manager=CallbackManager([\n", + " StreamingStdOutCallbackHandler()]),\n", + " verbose=True,\n", + " temperature=0.2,\n", + " max_tokens=150\n", + ")\n", + "\n", + "# use the LLM Chain to create a question creation chain\n", + "question_generator = LLMChain(\n", + " llm=llm,\n", + " prompt=condense_question_prompt\n", + ")\n", + "\n", + "# use the streaming LLM to create a question answering chain\n", + "doc_chain = load_qa_chain(\n", + " llm=streaming_llm,\n", + " chain_type=\"stuff\",\n", + " prompt=qa_prompt\n", + ")\n", + "\n", + "\n", + "chatbot = ConversationalRetrievalChain(\n", + " retriever=vectorstore.as_retriever(),\n", + " combine_docs_chain=doc_chain,\n", + " question_generator=question_generator\n", + ")" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": { + "tags": [] + }, + "outputs": [], + "source": [ + "# create a chat history buffer\n", + "chat_history = []\n", + "\n", + "# gather user input for the first question to kick off the bot\n", + "question = input(\"Hi! What are you looking for today?\")\n", + "\n", + "# keep the bot running in a loop to simulate a conversation\n", + "while True:\n", + " result = chatbot(\n", + " {\"question\": question, \"chat_history\": chat_history}\n", + " )\n", + " print(\"\\n\")\n", + " chat_history.append((result[\"question\"], result[\"answer\"]))\n", + " question = input()" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "## Customize your chains for even better performance" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": { + "tags": [] + }, + "outputs": [], + "source": [ + "import json\n", + "\n", + "from langchain.schema import BaseRetriever\n", + "from langchain.vectorstores import VectorStore\n", + "from langchain.schema import Document\n", + "from pydantic import BaseModel\n", + "\n", + "\n", + "class RedisProductRetriever(BaseRetriever, BaseModel):\n", + " vectorstore: VectorStore\n", + "\n", + " class Config:\n", + " \n", + " arbitrary_types_allowed = True\n", + "\n", + " def combine_metadata(self, doc) -> str:\n", + " metadata = doc.metadata\n", + " return (\n", + " \"Item Name: \" + metadata[\"item_name\"] + \". \" +\n", + " \"Item Description: \" + metadata[\"bullet_point\"] + \". \" +\n", + " \"Item Keywords: \" + metadata[\"item_keywords\"] + \".\"\n", + " )\n", + "\n", + " def get_relevant_documents(self, query):\n", + " docs = []\n", + " for doc in self.vectorstore.similarity_search(query):\n", + " content = self.combine_metadata(doc)\n", + " docs.append(Document(\n", + " page_content=content,\n", + " metadata=doc.metadata\n", + " ))\n", + " return docs" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "### Setup ChatBot with new retriever" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": { + "tags": [] + }, + "outputs": [], + "source": [ + "redis_product_retriever = RedisProductRetriever(vectorstore=vectorstore)\n", + "\n", + "chatbot = ConversationalRetrievalChain(\n", + " retriever=redis_product_retriever,\n", + " combine_docs_chain=doc_chain,\n", + " question_generator=question_generator\n", + ")" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "### Retry" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": { + "tags": [] + }, + "outputs": [], + "source": [ + "# create a chat history buffer\n", + "chat_history = []\n", + "\n", + "# gather user input for the first question to kick off the bot\n", + "question = input(\"Hi! What are you looking for today?\")\n", + "\n", + "# keep the bot running in a loop to simulate a conversation\n", + "while True:\n", + " result = chatbot(\n", + " {\"question\": question, \"chat_history\": chat_history}\n", + " )\n", + " print(\"\\n\")\n", + " chat_history.append((result[\"question\"], result[\"answer\"]))\n", + " question = input()" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [] + } + ], + "metadata": { + "colab": { + "provenance": [] + }, + "kernelspec": { + "display_name": "Python 3 (ipykernel)", + "language": "python", + "name": "python3" + }, + "language_info": { + "codemirror_mode": { + "name": "ipython", + "version": 3 + }, + "file_extension": ".py", + "mimetype": "text/x-python", + "name": "python", + "nbconvert_exporter": "python", + "pygments_lexer": "ipython3", + "version": "3.11.6" + } + }, + "nbformat": 4, + "nbformat_minor": 4 +} diff --git a/TalentLLM-main/output.txt b/TalentLLM-main/output.txt new file mode 100644 index 0000000000000000000000000000000000000000..1a9a0d60cd05b8772d24af932e56d04c960e7baf --- /dev/null +++ b/TalentLLM-main/output.txt @@ -0,0 +1,174 @@ +Daniyal Khan - Portfolio +Education +Georgia Institute of Technology +Bachelors of Science in Computer Science +August 2020 - December 2023 + +Concentrations: Intelligence/AI and Systems and Architecture +High Honours +Relevant Coursework: + +Operating Systems +Artificial Intelligence +Advanced Algorithms and Data Structures +Robotics and Perception +Computer Architecture +Circuit Design Lab + + +Links + +GitHub: danikhan632 +My Resume: Resume +LinkedIn: Daniyal M Khan +Email: danikhan632@gmail.com + +Vulkan Backend for Triton : August 2023 + +Developed Vulkan Backend for OpenAI’s Triton to enable Vulkan compatible devices utilization. +Addressed Vulkan’s SPIR-V entry point and descriptor sets requirements differing from OpenCL’s SPIR-V. +Proposed a modular, scalable architecture for Vulkan compute pipeline, adapting to dynamic configurations. +Explored JIT compilation and code injection for Vulkan, adapting Intel’s extension approach. +Created a Clang/g++ backend for dynamic C++ compilation, designed an interface for Vulkan integration. +Investigated memory management solutions for Vulkan compute integration with PyTorch. +Utilized MoltenVK for macOS ARM64 architecture ensuring consistent compute capabilities. +Enhanced SPIRV conversion processes for better compatibility with SPIRV-Cross. +Encouraged community contributions through detailed documentation and active engagement on Discord. + + +Guidance API: June 2023 + +Developed the Guidance API, integrating advanced language model capabilities for enhanced text generation and processing. +Enabled efficient network calls to Guidance, harnessing the power of cutting-edge language models for users. +Introduced a comprehensive output structure, supporting multiple generations, selections, conditionals, and tool use. +Optimized system performance with smart seed-based generation caching, ensuring efficient token storage. +Laid groundwork for future compatibility with role-based chat models, expanding the API’s versatility. +Enhanced control over modern language models, offering a superior alternative to traditional prompting and chaining. +Utilized intuitive syntax based on Handlebars templating, ensuring a user-friendly experience. +Enabled real-time interactions with Playground-like streaming in Jupyter/VSCode Notebooks. +Seamlessly integrated with Hugging Face models, introducing features like guidance acceleration, token healing, and regex pattern guides. +Emphasized model performance and precision, ensuring high-quality outputs and adherence to desired formats. + +Link Link +import guidance + +# set the default language model used to execute guidance programs +guidance.llm = guidance.llms.TWGUI("http://127.0.0.1:5000") + +# define a guidance program that adapts a proverb +program = guidance("""Tweak this proverb to apply to model instructions instead. + +{{proverb}} +- {{book}} {{chapter}}:{{verse}} + +UPDATED +Where there is no guidance{{gen 'rewrite' stop="\\n-"}} +- GPT {{#select 'chapter'}}9{{or}}10{{or}}11{{/select}}:{{gen 'verse'}}""") + +# execute the program on a specific proverb +executed_program = program( + proverb="Where there is no guidance, a people falls,\nbut in an abundance of counselors there is safety.", + book="Proverbs", + chapter=11, + verse=14 +) + +AutoGPT-Alpaca-Trader June 2023 + +Innovative Plugin Development: Spearheaded the design and implementation of a cutting-edge AutoGPT plugin, seamlessly integrating the GPT-4 powered AutoGPT application with Alpaca Trading API to augment algorithmic trading strategies with advanced AI capabilities. +API Integration and Security: Expertly established secure and efficient connections to Alpaca’s Trading API, enabling robust trade execution, account management, and real-time data retrieval functionalities, while ensuring data integrity and compliance with industry best practices. +Enhanced Trade Management: Developed a comprehensive suite of tools for the automated placement, modification, and cancellation of diverse stock and ETF orders, including market, limit, and stop orders, resulting in a streamlined trading experience and improved operational efficiency. +Account and Portfolio Management: Implemented advanced features for real-time monitoring and management of user account details, portfolio positions, and transaction history, delivering a holistic view of financial assets and enhancing user decision-making. +Market Data and Risk Management: Provided traders with access to vital real-time and historical market data, including stock quotes and bar data, as well as corporate action insights, complemented by a robust paper trading environment for strategy testing and risk mitigation. + + +AutoGPT Messages: May 2023 + +Developed the AutoGPT plugin for iMessages, enabling seamless integration with AI-powered messaging across multiple platforms, ensuring user data privacy and security. +Implemented a Python server backend, allowing the plugin to operate universally while maintaining a dedicated Mac server for core functionalities. +Streamlined the installation process with cross-platform support, providing detailed instructions for Linux, Mac, Windows, and WSL environments. +Enhanced user experience by integrating with the iMessage API and providing options for public accessibility using tools like tunnelto and ngrok. +Designed a user-friendly interface with real-time notifications, customizable settings, and integration capabilities with other communication tools for comprehensive messaging solutions. + + +Github Page +AutoGPT Local Infrence Server: May 2023 + +Developed the Auto-GPT-Text-Gen-Plugin to enable users to fully customize prompts for integration with locally installed large language models (LLMs), facilitating a shift away from dependency on GPT-4 and GPT 3.5. +Implemented a robust connection to Text Generation WebUI, serving as an API gateway for various models, which streamlines the process of managing complex configurations and environment settings. +Provided comprehensive documentation and a step-by-step installation guide, ensuring users can effortlessly download, configure, and utilize the plugin with their specific text generation setup. +Integrated flexibility for model selection and the ability to tweak generation parameters such as top_p, top_k, and repetition_penalty through environmental variables, enhancing user control over text generation outcomes. +Encapsulated API interactions and prompt management within the TextGenPluginController class, laying the groundwork for potential future expansions to support multiple APIs, thereby ensuring long-term maintainability and scalability of the plugin. + +Github Page +iMessages API: May 2023 + +Developed a Flask-based API to interact with iMessage, enabling users to send and retrieve messages as well as fetch recent contacts, enhancing communication automation. +Implemented secure access to the API by creating a custom decorator function that validates API keys, ensuring secure and authenticated interactions. +Orchestrated background data synchronization using threading, allowing for real-time updates of messages while maintaining a responsive API service. +Integrated iMessage reader and AppleScript for seamless message sending and retrieval, showcasing strong cross-technology integration skills. +Designed a user-friendly setup process, including environment variable configuration and easy-to-follow instructions, improving the accessibility of the API for end users. + +Github Page + +BuzzOS: January 2023 +BuzzOS is an Operating System built for the Intel/AMD x86_64 architecture using assembly and Rust. The operating system includes a Graphical User Interface (GUI) and is designed to provide a complete user experience. +The operating system includes user space and a mechanism for user-level processes to perform system calls to the kernel. This allows users to run applications and perform various tasks on the system. +BuzzOS also includes drivers for various hardware components, including the keyboard, mouse, timer, disk, and Intel PIC 8259. These drivers enable a robust input experience and ensure that the operating system can communicate effectively with various hardware components. +In addition to the core operating system functionality, BuzzOS also includes a fully functional desktop interface with games and system apps. This interface provides users with a familiar and intuitive environment for interacting with the operating system. +Overall, BuzzOS is an impressive project that demonstrates the power and flexibility of modern operating systems. By leveraging assembly and Rust, the project was able to create a complete operating system with a GUI and a range of drivers and applications. This is a significant achievement and represents a valuable contribution to the field of operating systems. Github Page +Path-finding Robot: October 2022 + +Developed proficiency in Robotics and Computer Vision through implementing the Rapidly-exploring Random Tree (RRT) algorithm, enhancing path planning efficiency in autonomous robotic navigation. +Leveraged Computer Vision techniques to enable real-time object detection and environment mapping, optimizing robot’s perception and decision-making capabilities. +Designed and executed algorithms for image processing and feature extraction, significantly improving the accuracy of object recognition in varied lighting and environmental conditions. +Employed state-of-the-art machine learning models for image captioning, translating visual data into descriptive language, and enhancing human-robot interaction. +Demonstrated strong problem-solving skills in Robotics by handling exceptions such as VectorTimeoutException, ensuring seamless operation and reliability of robotic systems. + +Github Page + +Flutter Tower Defense Game: April 2022 +Designed and developed a tower defense game using the Flutter framework. + +Implemented game mechanics including tower placement, enemy spawning, and pathfinding using the Dart programming language. +Utilized Flutter’s built-in animation framework to create smooth and visually appealing animations for tower attacks and enemy movements. +Integrated Google Firebase for user authentication and cloud storage to save game progress and scores. +Takes advantage of Flutter’s cross-platform nature, allowing it to run on iOS, Android, Mac, Windows, Linux, and Web. +Collaborated with a team of developers and designers to ensure timely delivery and a high-quality end product. + +Github Page +You can play the game here. + +COVID Vaccine Tracker: February 2021 +The COVID Vaccine Tracker is a tool for predicting the progress of COVID-19 vaccinations across US states. It uses data from vaccine databases and factors in state population to estimate when each state will reach an 80% vaccination rate. The project was created in March of 2021 but could potentially be modified for use with the Delta variant of COVID-19. +The model used in the project is based on a logarithmic curve. It provided fairly accurate predictions until the 50% vaccination mark but did not accurately predict the curve going logarithmic at that point. Despite this limitation, the tool still provides valuable insights into the progress of vaccinations across different US states. +Github Page +https://github.com/danikhan632/tower_defense_game +Create C++ App: November 2022 +Create-Cpp-App is a Command Line Interface (CLI) tool that provides an npm-like experience for building C++ applications. The tool is designed to streamline the process of building C++ apps by automating many of the repetitive and time-consuming tasks that developers typically face. +The tool is built to be intuitive and user-friendly, and it generates makefiles and automatically updates CMake files for a fast and efficient development experience. This allows developers to focus on writing code instead of worrying about the build process. +Create-Cpp-App also includes a range of built-in testing, address sanitization, benchmarking, and other tools for building production-ready C++ applications. These tools are designed to help developers ensure that their code is of high quality and performance. +Overall, Create-Cpp-App is an innovative tool that helps simplify the process of building C++ applications. By providing an npm-like experience, the tool makes it easy for developers to get started with building C++ apps and reduces the time and effort required to build high-quality, production-ready applications. + +Github Page +Clean Up Crew: October 2022 +Clean Up Crew is a web application that serves as a platform for connecting small communities with local businesses. The application was built using Next.js, MongoDB, AWS S3, Google Maps API, and ReactJS. +The platform allows users to create and interact with posts in a given area. Users can post about community events, local businesses, and other topics related to their community. The application includes a sorting algorithm based on various factors such as location, user interaction, and other metrics to ensure that the most relevant content is displayed to users. +The project was developed by a team of programmers who participated in a programming competition. Over a period of 36 hours, the team worked on developing the application and implementing its various features. After the competition, the team was awarded 13th place out of 191 teams, which is a testament to their hard work and the effectiveness of the application they developed. +Overall, this project represents a valuable contribution to small communities looking to improve their localities and small businesses seeking new opportunities. The platform provides a means for these groups to connect and collaborate, and the sorting algorithm ensures that the most relevant content is displayed to users. By utilizing modern web technologies and APIs, the platform is able to provide a seamless and user-friendly experience for its users. +Self-Driving-Car: January 2021 +The Self-Driving Car project is a machine learning project that aims to simulate the behavior of a self-driving car using a Convolutional Neural Network (CNN) and computer vision techniques. The project involves constructing a virtual environment where a car can be driven autonomously using machine learning algorithms. +The CNN is used to determine the speed and angle of rotation of the simulated vehicle based on data obtained from a virtual camera. The camera captures images of the environment and feeds them into the CNN, which processes the data and outputs a prediction for the vehicle’s next move. The CNN is trained using a dataset of labeled images and their corresponding speed and steering angles. +To implement the CNN, the project utilizes a number of machine learning libraries, including Tensorflow, Keras, and NumPy. These libraries provide a range of tools for developing, training, and testing machine learning models, as well as tools for processing and analyzing large datasets. +The project also includes a testing environment where the performance of the self-driving car can be evaluated. This environment allows the user to adjust parameters such as the speed and complexity of the environment, and to observe how the car responds to different scenarios. +Overall, the Self-Driving Car project represents an exciting application of machine learning and computer vision techniques to the field of autonomous vehicles. By simulating the behavior of a self-driving car in a virtual environment, the project provides a safe and scalable platform for testing and developing new algorithms and techniques for autonomous driving. + +Github Page +Amazon-Shopping Clone: December 2020 +The Amazon Shopping Clone is a web application built using the MERN stack (MongoDB, Express, React, and Node.js) and Stripe API. It mimics the design and user interface of the Amazon.com website, allowing users to browse and purchase products in a familiar environment. +One of the key features of the application is its login system, which allows users to create accounts and securely store their personal and payment information. This information is stored using MongoDB, a NoSQL database that provides a flexible and scalable data storage solution. +In addition to the login system, the application also utilizes the Stripe API to handle transactions in a secure and scalable manner. Stripe is a popular payment processing platform that provides a wide range of features for online businesses, including secure payment processing, subscription management, and fraud detection. +To ensure a smooth and intuitive user experience, the application implements a design language that closely mimics that of the Amazon.com website. This includes a consistent color scheme, typography, and layout, as well as familiar user interface elements such as navigation menus, search bars, and product listings. +Overall, the Amazon Shopping Clone provides a robust and scalable platform for online shopping that combines the familiarity and convenience of Amazon.com with the security and scalability of modern web technologies. Github Page +You can access the live demo of the FakeBlock Shopping project here + diff --git a/TalentLLM-main/requirements.txt b/TalentLLM-main/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..0ec2e92390c3e714bae369211f879d3f191abb15 --- /dev/null +++ b/TalentLLM-main/requirements.txt @@ -0,0 +1,7 @@ +litellm==1.0.0 +requests +openai==1.0.0 +google-cloud-aiplatform +gdown +PyGithub +gspread \ No newline at end of file diff --git a/TalentLLM-main/results.py b/TalentLLM-main/results.py new file mode 100644 index 0000000000000000000000000000000000000000..a8c377ed56eb31d75571febb3226cc9766806df9 --- /dev/null +++ b/TalentLLM-main/results.py @@ -0,0 +1,56 @@ +import gspread +from Candidate import JobCandidate +from typing import List +# Authenticate with Google Sheets using a service account +sa = gspread.service_account(filename='service_creds.json') + +def writeToSheets(candidates: List[JobCandidate]): + sh = sa.open("Figma_swe") + new_sheet_title = "Results" # Change this to your desired sheet name + + # Check if the sheet already exists + try: + existing_wks = sh.worksheet(new_sheet_title) + except gspread.exceptions.WorksheetNotFound: + existing_wks = None + + # If the sheet exists, delete it + if existing_wks: + sh.del_worksheet(existing_wks) + new_wks = sh.add_worksheet(title=new_sheet_title, rows="100", cols="10") # Adjust rows and cols as needed + + data_to_write = [ + [ "Timestamp", "Name", "Email", "Resume Link", "Cover Letter", "LinkedIn", "GitHub", "Personal Website", "Visa Sponsorship", "Disability Status", "Ethnic Background", "Gender", "Military Service" ] + + ] + + for candidate in candidates: + data_row = [ + candidate.timestamp.strftime("%m/%d/%Y %H:%M:%S"), + candidate.name, + candidate.email, + candidate.resume_link, + candidate.cover_letter, + candidate.linkedin, + candidate.github_link, + candidate.personal_website_link, + candidate.visa_sponsorship, + candidate.disability_status, + candidate.ethnic_background, + candidate.gender, + candidate.military_service + ] + data_to_write.append(data_row) + + new_wks.update('A1', data_to_write) + + + print(f"Data written to '{new_sheet_title}' sheet.") + + + + + + + + diff --git a/TalentLLM-main/resume_conversation.py b/TalentLLM-main/resume_conversation.py new file mode 100644 index 0000000000000000000000000000000000000000..36e87e001c2518611c5aa5fe4e66da2f9163edf1 --- /dev/null +++ b/TalentLLM-main/resume_conversation.py @@ -0,0 +1,28 @@ +import openai; +import json, os,sys +from dotenv import load_dotenv +load_dotenv() +from litellm import completion +from mathpix import extract_text +import gradio + +def get_prompt(candidate, chat_history, question): + return ('Given the details of a candidate, the previous chat history, and a question, answer the question as if you are the candidate. Keep the answers short and to the point.\n' + + 'Candidate Details:\n\n' + str(candidate) + '\nEnd Candidate Details\n' + + 'Chat History:\n\n' + chat_history + '\nEnd Chat History\n' + + 'Question:\n\n' + question + '\nEnd Question\n') + +def chat_with_candidate(candidate, model = 'chat-bison'): + chat_history = '' + print('You are now chatting with ' + candidate.name + '. Type in your question or type QUIT to stop.') + while True: + print('User:') + question = input() + print() + if question.strip().upper() == 'QUIT': + break + prompt = get_prompt(candidate, chat_history, question) + messages = [{ 'content': prompt, 'role': 'user'}] + response = completion(model = model, messages = messages)['choices'][0]['message']['content'] + print('Response:\n' + response + '\n') + chat_history += 'User:\n' + question + '\nResponse:\n' + response \ No newline at end of file diff --git a/TalentLLM-main/resume_conversation_interactive.py b/TalentLLM-main/resume_conversation_interactive.py new file mode 100644 index 0000000000000000000000000000000000000000..bcfa9000f36fb8fd8e5a3fe03b3148da0bc34dd3 --- /dev/null +++ b/TalentLLM-main/resume_conversation_interactive.py @@ -0,0 +1,49 @@ +import openai; +import json, os,sys +from dotenv import load_dotenv +load_dotenv() +from litellm import completion +from mathpix import extract_text +import gradio as gr + +model = 'chat-bison' +resume_location = 'resume_pdfs/1BXAuw6f1rDF05P734y_O7K8fYwgDVZvV.pdf' +resume_mmd = extract_text((resume_location)) + +def get_prompt(resume_mmd, chat_history, question): + history = '' + for user, bot in chat_history: + history += 'User:\n' + user + '\nResponse:\n' + bot + ' ' + return ('Given the resmue of a candidate, the previous chat history, and a question, answer the question as if you are the candidate. Keep the answers short and to the point.\n' + + 'Resume:\n\n' + resume_mmd + '\nEnd Resume\n' + + 'Chat History:\n\n' + history + '\nEnd Chat History\n' + + 'Question:\n\n' + question + '\nEnd Question\n') + +def inference(message, history, model = 'gpt-3.5-turbo'): + try: + flattened_history = [item for sublist in history for item in sublist] + full_message = " ".join(flattened_history + [message]) + messages_litellm = [{"role": "user", "content": get_prompt(resume_mmd, history, message)}] + partial_message = "" + for chunk in completion(model=model, + messages=messages_litellm, + stream=True): + if 'content' in chunk['choices'][0]['delta']: + partial_message += chunk['choices'][0]['delta']['content'] + yield partial_message + except Exception as e: + print("Exception encountered:", str(e)) + yield f"An Error occured please 'Clear' the error and try your question again" + +gr.ChatInterface( + inference, + chatbot=gr.Chatbot(height=400), + textbox=gr.Textbox(placeholder="Enter text here...", container=False, scale=5), + description=f""" + You are chatting with a resume.""", + title="Chat with Resume", + examples=["Introduce yourself."], + retry_btn="Retry", + undo_btn="Undo", + clear_btn="Clear", +).queue().launch(share = True) \ No newline at end of file diff --git a/TalentLLM-main/resume_mmds/Zahmad34@gatech.edu.pdf.mmd b/TalentLLM-main/resume_mmds/Zahmad34@gatech.edu.pdf.mmd new file mode 100644 index 0000000000000000000000000000000000000000..97dbd2207c072101a02aba83deade26395417aa8 --- /dev/null +++ b/TalentLLM-main/resume_mmds/Zahmad34@gatech.edu.pdf.mmd @@ -0,0 +1,95 @@ +Zaeem Ahmad + +C: 916-300-3117, E: Zahmad34@gatech.edu + +Nationality: USA + +EDUCATION + +$2021-2025$ + +Georgia Institute of technology + +BS, Major: Computer Science + +Atlanta, Georgia + +$2017-2020$ + +California Virtual Academy @ Sutter + +EXPERIENCE + +Honors: Dean's Honor List for 4 consecutive semesters; CGPA 3.96 / 4.00 + +Roseville, California + +Aug'23-Present VIP Undergraduate Researcher, Georgia Institute of Technology + +Atlanta, Georgia + +- Developed a robust object-oriented programming (OOP) framework with multiple classes and instance methods for substation component modeling, enhancing flexibility and scalability. + +- Integrated communication protocols to ensure seamless data exchange among diverse components, leading to improved efficiency within smart grid systems. + +- Conducted comprehensive research on component variables, providing valuable insights that significantly improved the accuracy of co-simulation models and informed data-driven decision-making. + +- Implemented co-simulation using the HELICS framework, ensuring accurate representation and interaction of components, thus contributing to the project's overall success. + +May'23-Present Undergraduate Researcher, Georgia Institute of Technology + +Atlanta, Georgia + +- Developed and implemented advanced Python-based computational models for cardiac electrophysiology, leading to a $30 \%$ improvement in arrhythmia understanding compared to existing models. + +- Successfully translated complex MATLAB algorithms to Python, resulting in a $20 \%$ reduction in computation time while maintaining accuracy and reliability. + +- Integrated high-throughput patch clamp data into the computational models to enhance parameter estimation precision. + +- Collaborated in the creation of a novel hybrid algorithm, reducing the need for initial assumptions in parameter estimation by $40 \%$, thus significantly enhancing accuracy. + +Jan'23-Aug'23 Big Data Big Impact CS Intern, Georgia Institute of Technology Atlanta, Georgia + +- Spearheaded a cross-functional team effort to design and develop the Hurricane Cost Projector, a cutting-edge disaster impact prediction tool. + +- Employed HTML, CSS, Python, and JavaScript to craft an intuitive and visually appealing front-end interface, ensuring a user-friendly and engaging experience. + +- Enhanced the tool's effectiveness through the implementation of data visualization features, facilitating customer comprehension and interaction. + +- Thrived in a dynamic, fast-paced environment, collaborating seamlessly within cross-functional teams to deliver impactful and innovative solutions to drive project success. + +May'22-Aug'23 BluEast - Technical Intern + +Lahore, Pakistan + +- Assisted with PCB design \& embedded programming of an iOS \& Android compatible App, "Autopilot", that is built using JavaScript (Frontend \& Backend frameworks), Python, Ruby, PHP, \& C++ + +- Spearheaded optimization process for algorithm that employs Mevris AI to tweak thermostat of DC inverter air conditioner based on metrics to ensure thermal comfort \& reduce energy consumption by $80 \%$ + +- Developed an automation scheme for test condition parameters that saved configuration time on the system $\&$ resulted in an overall $50 \%$ reduction in validation testing time. + +Sep'20-Aug'21 Orient Group of Companies - Business Analyst + +Lahore, Pakistan + +- Digitized sales network through 'Orient Dost' app that enabled $76+$ team members \& 100+ distributors spread across 16 territories to track performance in real time which led to a $20 \%$ increase in sales. + +- Automated dispenser production line to enhance efficiency \& replaced steel tanks with aluminum to improve hygienic standards, reduce production cost by $4 \%$, \& achieve $100 \%$ annualized increase in sales . + +- Led team of 4 to implement employment 'Training Need Program' in SAP ERP to improve productivity of 67 middle management employees who completed 6 different training programs as part of this initiative. + +\title{ +EXTRA-CURRICULAR +} + +- National Champion at National Tournament of Young Math, National STEM Educational Foundation Tournament of Young Mathematicians 2019. + +- Contestant at International Earth Science Olympiad (2020 event canceled due to COVID-19). + +- National Champion at National Earth Science Olympiad, National STEM Educational Foundation 2019-2020. + +SKILLS + +- Programming Languages - C++, Python, Java, JavaScript, HTML, React.JS, React Native, MATLAB, CSS, Figma, GitHub, Git + +- $\quad$ Languages - English (Fluent), Urdu (Native), Hindi (Fluent), Punjabi (Intermediate) \ No newline at end of file diff --git a/TalentLLM-main/resume_mmds/azishaqui@gmail.com.pdf.mmd b/TalentLLM-main/resume_mmds/azishaqui@gmail.com.pdf.mmd new file mode 100644 index 0000000000000000000000000000000000000000..01c8d8570865502f089128804c09971bfc1a8589 --- /dev/null +++ b/TalentLLM-main/resume_mmds/azishaqui@gmail.com.pdf.mmd @@ -0,0 +1,83 @@ +\title{ +Azeez Ishaqui +} + +- Gmail — azishaqui@gmail.com $・$ LinkedIn $\cdot$678-313-6772 $\cdot$ U.S. Citizen + +\section*{EDUCATION} + +Georgia Institute of Technology - Atlanta, GA + +M.S. in Machine Learning, Current Masters student, BS/MS program — GPA: 4.0/4.0 + +Georgia Institute of Technology - Atlanta, GA + +B.S. in Computer Science (Artificial Intelligence) — GPA: 4.0/4.0 + +Technical Skills: Python, PyTorch, Java, C/C++, Scala, AWS (EC2/STS), Terraform, Docker, SQL, HTML/JS/CSS, Assembly + +Languages: French (Advanced), English (Native) + +Relevant Coursework: Machine Learning, Deep Learning, Software Development, Data Structures and Design/Analysis of Algorithms, Probability/Statistics, Linear Algebra, Combinatorics, Discrete Mathematics, Multivariable Calculus, Introductory AI + +\section*{WORK EXPERIENCE} + +KPMG CIO Advisory + +June 2023 - August 2023 + +- Constructed an AI chatbot with full-stack skills including HTML/JS/CSS embedded onto Unqork to manage client database + +- Leveraged ML tools like Scikit, NumPy, and Pandas to improve Great American Health user experience and customer retention + +- Implemented a scalable deployment of the chatbot with a Flask API using NGINX reverse proxy and Gunicorn on an $\underline{\text { AWS }}$ EC2 instance, integrated with OpenAI GPT-3 chat functionality on a secure environment with proper user access control + +- Conducted data analysis, prepared reports, and utilized Excel and data visualization tools to assist executive decision-making + +- Analyzed comprehensive migration strategies and roadmaps, including assessment of on-premises healthcare infrastructure, application analysis, and cost optimization plans, resulting in efficient and cost-effective cloud deployments within $\underline{\text { GCP }}$ + +ASI (Aggregate Singularity Industries) Cloud/Devops Engineer + +June 2022 - August 2022 + +- Developed cloud-based AWS services, API integrations with Snyk for compromised repositories security, bucket pruning for performance/cost optimization, and an EC2 discovery service with instance tagging and Prometheus scrape configuration + +- Created a Slack-integrated service for company website traffic monitoring, containerized on Docker, configured with $\underline{\text { Grafana }}$ and Prometheus to initiate error threshold alerting for company-wide server maintenance, critical for users and employees alike + +- Provisioned and managed resources on Terraform, including load balancers, access keys, and customized port configurations for efficient traffic handling, while maintaining domain server privacy for developers + +- Utilized EC2/STS services and Terraform resource provisioning to administer an AWS replication service that safeguarded server functionality during internal network failures and complete outages + +Assurant Inc. Team Project Lead \& ML System Designer + +January 2023 - Present + +- Automated data ingestion scraping FRED macro-economics handled on Python for a housing price prediction model with metro-Atlanta data with pre-processing techniques and supervised learning including multivariate time-series analysis and $\underline{\mathbf{K}-}$ nearest neighbors to minimize loss and reduce overfitting + +- Designed and co-implemented an $\underline{\text { ML pipeline with Assurant engineers focusing on precision and efficiency within AWS }}$ + +Cognira (Retail/Analytics Software Company) Software Development Intern + +June 2021 - August 2021 + +- Administered a web server which ingested client data and performed several forms of data optimization analysis using Scala + +- Performed server testing using an external software client and Docker containers for application isolation and portability + +Eye Consultants of Atlanta Data Analytics Intern - Eye Consultants of Atlanta + +January 2021 - February 2021 + +\section*{LEADERSHIP/PUBLICATIONS} + +Founder of "Charity 4 All", a non-profit organization - Charity 4 All Solo Project + +March 2017 - Present + +- Publications: Atlanta Journal Constitution - $\underline{\mathrm{AJC}}$ + +- Fundraised over $\$ 2500$ through GoFundMe, leveraged social media and personal network to spread awareness of the nonprofit, used surplus money raised to provide pencils, notebooks, textbooks, to support over twenty students' education + +- Facilitated international logistics to ship and hand-deliver 2000+ masks to several different Atlanta Homeless Shelters + +- Invigorated local economy by providing carts and sewing machines to needy vendors and widows, annually distributed food staples and durable goods during Ramadan to impoverished families with limited access after fasting \ No newline at end of file diff --git a/TalentLLM-main/resume_mmds/dan.hus120@gmail.com.pdf.mmd b/TalentLLM-main/resume_mmds/dan.hus120@gmail.com.pdf.mmd new file mode 100644 index 0000000000000000000000000000000000000000..bb24eee8d24ab88b5594680cfbfb189426662781 --- /dev/null +++ b/TalentLLM-main/resume_mmds/dan.hus120@gmail.com.pdf.mmd @@ -0,0 +1,107 @@ +\title{ +Daniyal Hussain +} + +New York, NY | 347-658-6214 | dh3144@nyu.edu | + +www.daniy.al | linkedin.com/in/hussain-daniyal/ | github.com/tuffstuff9 | + +EDUCATION + +New York University, College of Arts and Sciences, New York, NY + +Bachelor of Arts in Computer Science and Data Science + +Graduated: May 2023 + +Cumulative GPA: 3.5 + +Technical Skills + +Languages: TypeScript, JavaScript, C\#, Java, Python, R, HTML/CSS, SQL + +Frameworks/Technologies: React, Next.js, PlanetScale, TensorFlow, PyTorch, Hadoop, Maven, WinUI 3, Git, Jira, Figma, REST API, Postman, Node.js, Agile/Scrum, AWS, Azure, Google Cloud, Jest, Mocha, Chai, Docker + +\section*{Professional Experience} + +Deloitte + +June 2022 - August 2022 + +Software Engineer Intern + +Remote + +- Engineered SPA and SQL database for Ministry of Mining's user/business accounts with robust authentication + +- Implemented page wireframes by designing front-end using JavaScript and OutSystems with optimized UX + +- Delivered functional prototypes to enable client evaluations and feedback on user experience and design aspects + +Berry Mount + +June 2021 - January 2022 + +Software Engineer + +Remote + +- Overhauled company ERP and CRM software from excel spreadsheet/FocusERP hybrid-system to Odoo software + +- Automated and trained team on invoice tracking, payment deposits, inventory management, vendor/client reports + +- Integrated missing invoice tracker and mitigated inventory loss, restoring $\$ 4,000-\$ 10,000$ in business transactions + +USME Supply Lines Trading + +June 2020 - August 2020 + +Business Analyst Intern + +Los Angeles, CA + +- Constructed an internal program to produce pricing quotes, increasing quote frequency from weekly to biweekly + +- Sourced new freight partners to reduce shipment costs, increasing weekly operational profitability by $\$ 1,000-\$ 3,000$ + +- Conducted cost-benefit analyses for business expansion, leading to the successful launch of two new products + +University of California Los Angeles $\quad$ June 2018 - July 2018 + +Research Assistant + +Los Angeles, $C A$ + +- Streamlined manual data retrieval process from NIST to study effects of cryogenic cooling on hydrogen fuel + +- Scraped database using Python and VBA with variable input parameters to generate error reports for models + +Projects + +wandy.ai - SaaS AI Model Tuner | TypeScript, React, Next.js, ClerkJS, PlanetScale, Drizzle, Zod, Stripe + +- Engineered full-stack solution for real-time AI model tuning \& comparison, prompt generator, and training on documents + +SlideResume.com - SaaS cover letter generator | TypeScript, React, Next.js, ClerkJS, PlanetScale, Drizzle, Zod, Stripe + +- Produced a user-centric SaaS web app which automates personalized cover letter generation from uploaded resume + +Leetcode AI Assistant - Chrome Extension | JavaScript, React, Chrome MV3 + +- Designed and built a Chrome extension which acts as an automated flashcard system for LeetCode, with AI assistance + +- Leveraged latest Chrome APIs (MV3) to incorporate efficient DOM manipulation and optimize performance + +KeyboardHookLite $\mid$ C\# + +- Developed a low-level keyboard hook library for modern UI frameworks, with unmanaged code disposal + +- Published on Nuget with over $3000+$ user downloads and on GitHub with detailed documentation + +NYU Course Tracker | JavaScript, Puppeteer, Telegram + +- Utilized headless browser to scrape NYU course status and alert users in real-time for their class openings + +- Bypassed session expiry using stealth features to avoid detection by randomly manipulating user agent + +Other Projects: nextjs-pdf-parser, PlanetScale + Drizzle Boilerplate, OCR AI Assistant (Chrome Extension), Visualizer.coffee Shot Downloader (Website Scraper), ChatGPT Clipper (Chrome Extension), React Blackjack (Web App) \ No newline at end of file diff --git a/TalentLLM-main/resume_mmds/danikhan632@gmail.com.pdf.mmd b/TalentLLM-main/resume_mmds/danikhan632@gmail.com.pdf.mmd new file mode 100644 index 0000000000000000000000000000000000000000..513a01568c4fc558374963657d384b85e70c3f74 --- /dev/null +++ b/TalentLLM-main/resume_mmds/danikhan632@gmail.com.pdf.mmd @@ -0,0 +1,133 @@ +\title{ +Daniyal Khan +} + +O danikhan632 | in daniyalmkhan/ | A + +danikhan632@gmail.com | US Citizen | www.daniyalkhan.dev/ + +EDUCATION + +Georgia Institute of Technology + +B.S. Computer Science + +Expected December 2023 + +High Honors + +Concentrations: Intelligence/AI and Systems and Architecture + +Relevant Coursework: Agile Development, Artificial Intelligence, Advanced Algorithms and + +Data Structures, Robotics and Perception, Computer Architecture, Circuit Design Lab + +EXPERIENCE + +OpenAI + +July 2023- + +Open Source Contributor SWE + +Triton Project + +- Implemented an LLVM backend interface using $\mathrm{C}++$ to facilitate the conversion between different LLVM dialects, bringing Triton to Apple Silicon Macs. + +- Combined $\mathrm{C}++$ expertise with SPIRV-Cross to create an integrated cross-compiling toolchain, enabling efficient conversion from diverse LLVM-Dialects to Metal Shader Language + +- Architected and developed a JIT compiler for GPU Metal kernels within Triton, harnessing the power of C++ and SPIRVCross, resulting in a remarkable $70 \%+$ speed increase in large matrix multiplication tasks on Apple Silicon Macs. + +Microsoft- Guidance Project + +May 2023 - July 2023 + +Open Source Contributor SWE + +- Architected and implemented LLM support system with more efficient memory and disk-swapping for larger LLMS + +- Harnessing the power of Token and Regex processors, led the orchestration of hundreds of LLMs to generate structured, schemaful data, boosting accuracy from $63.04 \%$ to $76.01 \%$. + +- Transformed the token generation process by offloading operations to integrated REST API. + +NCR Software Engineering + +May 2022 - August 2022 + +Software Engineering Intern + +Atlanta, Georgia + +- Led the creation of an internal debugging tool, facilitating real-time monitoring and management of MQTT messages + +- Designed and implemented a dynamic frontend using React, deeply integrating TypeScript and Redux to ensure a seamless user experience and efficient state management. + +- Mastered the intricacies of SQL to ensure optimal logging, storage, and retrieval of MQTT messages, enhancing system responsiveness and reliability. + +- Pioneered a custom TreeSet data structure, optimizing data modification and retrieval processes + +PROJECTS + +Auto-GPT + +2023 + +Machine Learning Engineer + +AI-based Financial Trading System + +- Architected an AI system enabling agents to autonomously trade stocks by analyzing real-time financial data and trends. + +- Seamlessly integrated AI with Apple's iMessage system, empowering the agent to craft contextually accurate text responses + +- Pioneered an AI-driven task management system via Todoist integration, resulting in notable increases in user productivity. + +iMessage API + +Software Developer + +Chatbot Integration Platform + +- Developed a Flask Python API to bridge Apple's iMessage with external devices, enabling applications such as chatbots. + +- Automated iCloud contact synchronization via the API, providing real-time contact updates to all users. + +BuzzOS + +2023 + +System Architect + +Monolithic x86 Operating System + +- Spearheaded the architecture of a Rust-based Kernel Driver tailored for efficient graphics rendering. + +- Designed and integrated user libraries that facilitated system calls from userspace processes. + +- Broadened hardware support by crafting specific drivers for vital components, including keyboard, mouse, and timer. + +Clean Up Crew + +Full-Stack Developer + +Community Improvement Platform + +- Conceptualized and developed a crowdsourcing platform, bridging communities with contractors for targeted improvements. + +- Employed a robust tech stack, integrating Next.js and MongoDB to design a fluid and responsive UI + +- Pioneered an intuitive posting mechanism, strategically prioritizing community issues based on user engagement metrics and location data. + +Create-Cpp-App: $\mathbf{n p m}$ for $\mathbf{C}++$ + +System Architect + +- Build an inuitive CLI tool for building C ++ apps with an npm experience + +- generates makeg fles and automatically updates CMake files for fast developer experience + +- includes built-in testing, address sanitization, benchmarking and other tools for building production-ready $\mathrm{C}++$ apps + +Programming languages: Rust, C++, TypeScript, Go, C++, Dart, Python, Java, C\#, SQL, Bash, JavaScript, HTML, CSS + +Frameworks Software: Flutter, Axios, Flask, Docker, LLVM, CUDA, Pytorch, React, Springboot, Maven, React Native \ No newline at end of file diff --git a/TalentLLM-main/resume_mmds/himothy2110@gmail.com.pdf.mmd b/TalentLLM-main/resume_mmds/himothy2110@gmail.com.pdf.mmd new file mode 100644 index 0000000000000000000000000000000000000000..b7e81165e343a36cfc3b18376d6844e62efec2fb --- /dev/null +++ b/TalentLLM-main/resume_mmds/himothy2110@gmail.com.pdf.mmd @@ -0,0 +1,85 @@ +\title{ +Angelina Lee +} + +415-812-3329 | $\underline{\text { AngelinaTheDev@gmail.com | linkedin.com/in/AngelinaLee | github.com/AngelinaTheDev }}$ + +\section*{EDUCATION} + +University of California Berkeley + +B.S. in Computer Science +Berkley, California + +Aug. 2014 - May 2018 + +EXPERIENCE + +Senior Full Stack Engineer + +Instagram +Sep. 2020 - Present + +San Francisco, $C A$ + +- Spearheaded the development of a revolutionary news feed infrastructure utilizing React for AI on Blockchain, transforming user interaction and content delivery. + +- Pioneered the HyperQuantum Feed Algorithm using server-side React Larceny AI, optimizing web app performance and resolving big data pipeline issues in record time. + +- Solved complex scalability issues by employing the use of useless architectures and microservices, orchestrated through the use of Quantum Kubernetes (QK8s). + +- Led a dynamic team of 6 engineers in a covert operation to mine Ethereum on company servers, generating unprecedented profits and resources for $R \& D$. + +\begin{tabular}{lr} +Senior Full Stack Engineer & May 2018 - August 2020 \\ +Zillow & Seattle, WA +\end{tabular} + +- Implemented advanced AI optimization techniques to the GraphQL API, achieving a $69 \%$ improvement in page load times and significantly enhancing user experience. + +- Conducted user research to assess the impact of gamified elements in property listings, resulting in valuable insights for future user interface enhancements. + +- Played a vital role in the development process by contributing over 50,000 lines of clean, maintainable code to an established codebase via Git. + +- Organized and facilitated team-building activities, including a company-wide potato sack race, which fostered a sense of camaraderie and improved team cohesion. + +- Embraced the critical role of Team Coffee Alchemist, ensuring the team of 6 was fully caffeinated with ultra-rare Antarctican coffee beans, ground to an optimal 14nm particle size using proprietary grinding technology. + +Software Engineer Intern + +Microsoft +May 2016 - Aug 2016 + +New York, NY + +- Developed and optimized SQL queries and stored procedures, resulting in a $40 \%$ improvement in database performance for a critical customer-facing application. + +- Contributed to the backend development of a new feature using Spring Boot, adhering to microservices architecture principles for enhanced modularity and scalability. + +- Spearheaded initiative that enabled the dispersion of acquired immunodeficiency syndrome across the platform team + +- Collaborated with cross-functional teams to understand requirements and deliver high-quality software solutions within agile development sprints. + +\section*{Projects} + +Gitlytics | Python, Flask, React, PostgreSQL, Docker + +June 2020 - Present + +- Developed a full-stack web application using with Flask serving a REST API with React as the frontend + +- Implemented GitHub OAuth to get data from user's repositories + +- Visualized GitHub data to show collaboration + +- Used Celery and Redis for asynchronous tasks + +Technical Skills + +Languages: Java, Python, C/C++, SQL (Postgres), JavaScript, HTML/CSS, R + +Frameworks: React, Node.js, Flask, JUnit, WordPress, Material-UI, FastAPI + +Developer Tools: Git, Docker, TravisCI, Google Cloud Platform, VS Code, Visual Studio, PyCharm, IntelliJ, Eclipse + +Libraries: pandas, NumPy, Matplotlib \ No newline at end of file diff --git a/TalentLLM-main/resume_mmds/jamesryzhkov@gmail.com.pdf.mmd b/TalentLLM-main/resume_mmds/jamesryzhkov@gmail.com.pdf.mmd new file mode 100644 index 0000000000000000000000000000000000000000..053746b6a35640167852446ac6bba3e7f3751514 --- /dev/null +++ b/TalentLLM-main/resume_mmds/jamesryzhkov@gmail.com.pdf.mmd @@ -0,0 +1,101 @@ +Programming Languages and Frameworks: Java, Next.js, React.js, JavaScript, HTML \& CSS, Node.js, Python, C, C\#, C++, SpringBoot, PostgreSQL, SQL, MongoDB, Swift, SwiftUI, Bun, Git, GCP + +Natural Languages: English, Belorussian, Russian, Ukrainian + +EDUCATION + +Georgia Institute of Technology + +BS Computer Science | Systems Architecture, Internetworks, GPA: College 3.64 + +Southern Polytechnic University + +BS Computer Engineering |Honors College, GPA: 3.7 + +Relevant Courses: + +Data Structures \& Algorithms, Objects \& Design, Design \& Analysis-Algorithms + +Computer Organization \& Programming, Systems and Networks, Design of Operating Systems + +Database Systems, Advanced/High Perform Computer Architecture, Computer Networking, Processor Design + +EXPERIENCE + +Bezalel Studio (B) + +Lead Developer 83 Director + +- Manage a team to offer website services, branding, SEO \& SMM, Mobile and Computer Applications. + +- Handle discussions with clients to understand requirements and business specifications while guiding the team. + +- Designed and implemented a unified payment portal that integrates Zelle, Paypal, ApplePay, Credit Card, and Venmo, reducing third-party transaction costs by $76 \%$. + +Georgia Tech VIP Program | Stadium Internet of People and Things + +Team Lead - App Developer and Researcher fanplay.tech + +- VIP team targeted at enhancing Georgia Tech sports fans' game-day experience and ensuring their security by enabling their mobile devices to access innovative infotainment and venue-related information. + +- Research focused on integration and development for the iOS Application. + +BMI Inc. + +Information Technology Web Developer + +- Responsible for the electronic necessities in running the non-profit 501 (c)3. + +- Created Dynamic systems for donation processing, custom email marketing campaigns, maintained servers. + +- Increased online donations by $83 \%$ by redesigning the donation portal and improving $\mathrm{U} / \mathrm{X}$. + +KJ Web \& Productions + +Founding Director + +Cumming, GA + +April 2015 - September 2018 + +- Created 3-D Animations and Graphics for Jumbotron applications. + +- Managed to increase the price per second from $\$ 5$ to $\$ 18$ per second by introducing modular reuse design within Cinema4D sets. + +\title{ +PROJECTS +} + +Adapt Booking React, PostgreSQL, SpringBoot, Java + +book.chanceyouth.org + +Created a web based application that allowed event managers to have a system that could dynamically adjust during registration. Currently, working on bots to automate the check-in process. + +NFC ATM React, Node.js + +Application that runs on an ATM machine that is configured to NFC 13.5GHz. Point system used for children events to imitate an economy with working VISA cards. + +Tale - Purposeful Management Express Node.js Framework, HTML, CSS, JavaScript + +Simple yet powerful team management web-app that focused on binary emulated sub-team collaboration. + +LEADERSHIP \& ACTIVITIES + +Chance Youth @chance.youth, Ukraine, Poland, Belarus, United States + +chanceyouth.org + +Co-founded Project Chance to empower young people that have a heart for compassion, to travel overseas to conduct camps for orphans. The project focuses on sharing the Love of Christ with the unfortunate outcasts of society. This positive movement has grown to help bring financial, emotional, and spiritual comfort to hundreds of people year-round. + +GT iOS club SwiftUI, Firebase, Auth, Firestore, Realtime + +Working on app "GT Collaboration" + +Assignments: $\bullet$ Integrate 2FA $\bullet$ Smart Search $\bullet$ Swift Charts $\bullet$ UI Design (Figma) + +Startup LAB Georgia Tech Evidence-Based Entrepreneurship + +create-x.gatech.edu + +GT WebDev club Active member developing modern-web applications / (Spring Project - "Punchshot") \ No newline at end of file diff --git a/TalentLLM-main/resume_mmds/mr.zeyadalsakhi@gmail.com.pdf.mmd b/TalentLLM-main/resume_mmds/mr.zeyadalsakhi@gmail.com.pdf.mmd new file mode 100644 index 0000000000000000000000000000000000000000..32578464991411010b05f8b97e3744b5214c14c9 --- /dev/null +++ b/TalentLLM-main/resume_mmds/mr.zeyadalsakhi@gmail.com.pdf.mmd @@ -0,0 +1,109 @@ +\title{ +Zeyad Al Sakhi +} + +( github.com/zeyadtmi | in linkedin.com/in/zeyadalsakhi | $\boldsymbol{\text { zsakhi3@gatech.edu | }}$ +1 4049553707| Atlanta, GA EDUCATION + +Georgia Institute of Technology + +B.S. Computer Science - Bahrain Crown Prince International Scholarship + +Aug 2021 - May 2024 + +Concentrations: Computer Systems Architecture and Information Internetworks + +TeChnical Skills + +Software languages: $\mathrm{C}, \mathrm{C}++$, Python, Java, Assembly, VHDL + +Web Technologies: React, HTML, CSS, JavaScript + +ML/AI: Numpy, Pandas, ScikitLearn, Matplotlib + +Other: MySQL, Git, LateX, Android Studio, Agile, Jira + +Relevant Coursework: Advanced Computer Architecture, Processor Design, Operating Systems, Machine Learning, Artificial Intelligence, Advanced Algorithms and Data Structures, Networking, OOP, Algorithmic Design, Databases + +EXPERIENCE + +Investcorp + +Security Engineer + +June 2023 - August 2023 + +Bahrain + +- Developed an Outlook Metadata Anomaly Detector using Azure Synapse Analytics and PySpark for data processing and analysis, flagging suspicious email activities. + +- Analyzed 500,000+ email meta entries, identifying $0.2 \%$ as potential threats based on features like excessive BCC to external domains, irregular email timings, and unusual recipient interactions. + +- Conducted broad scanning and vulnerability analysis, investigating source and destination IPs and port numbers, improving vulnerability detection by $25 \%$ and enhancing network security. + +- Designed a Ransomware Incident Handling Playbook, outlining procedural guidelines and response strategies to efficiently address potential ransomware attacks, reducing response time by $40 \%$. + +Bapco Energies + +May 2023 - June 2023 + +Cyber Security Trainee + +Awali, Southern Governorate, Bahrain + +- Managed SOC incident handling using Microsoft Sentinel and Microsoft Defender Security, streamlining threat detection and mitigation processes, resulting in a $20 \%$ faster response to incidents. + +- Developed a web scraper to extract security-related keywords from Google articles, enhancing employee awareness and reducing security-related incidents by $15 \%$. + +- Employed Azure Sentinel and KQL for real-time data extraction and analysis from security logs, enabling prompt threat detection and response, analyzing 100+ events daily. + +Georgia Tech Vertically Integrated Program + +Machine Learning Research Team Lead + +Aug 2022 - Present + +Atlanta, Georgia + +- Led the largest research subteam, engineering and deploying SVM, K-Nearest Neighbor, and YOLOv8 CNN models, achieving a 95\% accuracy in classifying microscopic 3D print images into defective and non-defective categories. + +- Employed feature extraction techniques including Histogram of Oriented Gradients (HOG), enhancing model performance and reducing false negatives by $20 \%$, critical for early defect detection. + +- Conducted rigorous iterative testing, model tuning, and data augmentation strategies, improving classification accuracy by $15 \%$, significantly elevating the quality control standards in $3 \mathrm{D}$ printing processes. + +PROJECTS + +\section*{CryptoLearn} + +- Developed a live crypto and stocks tracker using Binance and Yahoo Finance APIs, providing real-time data on prices across $10+$ currencies, tracking $200+$ daily trades, and monitoring $50+$ assets' market metrics. + +- Created a financial dashboard with matplotlib and CryptoCompare API, analyzing 100+ cryptocurrencies' returns over various periods, and rendering data visualizations for enhanced investment insights. + +\section*{DayMaker} + +- Developed DayMaker, streamlining event and deadline tracking by processing uploaded documents with OCR and NLP, and syncing extracted dates to Google Calendar, automating 500+ event entries. + +- Implemented secure authentication and authorization for Google Calendar and NLP Google Cloud APIs access, enhancing system security. + +- Orchestrated OAuth 2.0 Protocol for secure Google Calendar API requests, ensuring accurate data synchronization. + +- Achieved 1st place in the Google Cloud Category at HackTX hackathon. + +LEADERSHIP + +Undegraduate Research Ambassadors + +Computer Science Research Ambassador + +August 2023 - Present + +Atlanta, Georgia + +- Spearheaded undergraduate research initiatives, boosting engagement by $30 \%$ with the involvement of over 200 students. + +- Offered personalized guidance through one-on-one sessions with $50+$ freshmen and sophomores, aiding their integration into $20+$ ongoing research projects. + +Bahrain Crown Prince International Scholarship Program Alumni Association $\quad$ Jul 2021 - Present Academic Mentor + +- Mentored 30+ candidates in the Crown Prince International Scholarship Program (CPISP), aiding in their academic and career advancements. + +- Conducted 50+ individual remote consultations, providing tailored advice and fostering a supportive environment. \ No newline at end of file diff --git a/TalentLLM-main/resume_mmds/nhossain34@gatech.edu.pdf.mmd b/TalentLLM-main/resume_mmds/nhossain34@gatech.edu.pdf.mmd new file mode 100644 index 0000000000000000000000000000000000000000..f1f8935326edf26059812dbed2d690e6747b0cd9 --- /dev/null +++ b/TalentLLM-main/resume_mmds/nhossain34@gatech.edu.pdf.mmd @@ -0,0 +1,95 @@ +\title{ +Nabeel Hossain +} + +\author{ ++1701-740-0406nhossain34@gatech.edu linkedin.com/in/nabeel-hossain +} + +Education + +Georgia Institute of Technology + +Expected May 2025 + +Master of Science in Computer Science, Machine Learning + +Georgia Institute of Technology + +Expected May 2024 + +Bachelor of Science in Computer Science, Intelligence and System Architecture + +GPA: 4.0 + +\section*{Experience} + +Teaching Assistant, Georgia Tech College of Computing - Atlanta, GA + +- Debugged and graded assignments in C using GDB. Provided comprehensive feedback to significantly enhance students' understanding of systems level programming and problem solving strategy. + +- Facilitated lab sessions to review and answer questions on fundamental course topics using slides, example problems, and hands on demonstrations to over 40 students weekly. + +- Provided exceptional support and guidance to students in-person, virtually, and via discussion forums, promptly addressing their questions and concerns to create a highly conducive learning environment. + +- Improved efficiency in debugging, testing, and grading by a factor of 5 using python and Linux shell tools. + +Bits of Good, Georgia Tech College of Computing - Atlanta, GA + +Sept 2022 - Present + +- Developed, performed code reviews, and organized technical leadership for nonprofit development + +- Collaborated within a 9-member agile team structure, working closely with an Engineering Manager, a Product Manager, and a team of 4 developers to deliver high-quality projects + +- Directed project development by creating sprint tickets, prioritizing tasks and ensuring timely progress + +- Demonstrated strong technical expertise by reviewing pull requests, providing constructive feedback to fellow developers, and resolving merge conflicts promptly and effectively + +Projects + +Customer Churn Prediction + +- Predictive model for judging attrition rates for credit card customers given various features + +May 2023 - July 2023 + +- Collaborated with and coordinated tasks on 5-person team; utilized popular python data science and visualization libraries, such as NumPy, Pandas, PyPlot, and SKLearn + +- Conducted thorough data exploration, feature selection, and class balancing on dataset of 10,000 customers + +- Achieved 0.96 weighted f-1, 0.90 minority class recall after rigorous, documented model tuning and selection. + +Earthcraft + +Sept 2022 - Apr 2023 + +- Developed a NextJS 12 application backed by MongoDB and Azure Blob Storage, efficiently organizing a digital library of over 1,000 sustainable building standards to be taggable, searchable + +- Implemented a comprehensive solution for creating, searching, organizing, editing, and sharing standards, enabling clients to organize their information 3 times faster than before + +- Designed the innovative Report Builder Feature, allowing engineers to seamlessly generate project reports by combining multiple standards, images, and user notes. Accelerated productivity with one-link PDF sharing. + +- Utilized SWR for responsive UI updates and Server-Side Rendering for optimized loading speed. + +CleanUpCrew + +Apr 2020 - June 2020 + +- Developed a dynamic decentralized crowdfunding platform using NextJS, MongoDB, and Google Maps API. + +- Used aggregated geospatial data to design and implement a unique algorithm that can identify posts with the greatest potential community impact to prioritize to viewers + +- Crafted a sleek and responsive user interface, ensuring seamless user experience across various devices. + +- Coordinated frontend and backend tasks within a lean team of four, successfully delivering the project within an impressive 36-hour timeframe. + +Skills + +Languages: Python, Java, JavaScript, TypeScript, C, C++, HTML/CSS + +Frameworks: ReactJS, Node.js, Express.js, Next.js, Flask, SCSS, jQuery, Mongoose.js + +Technologies: SQL, MongoDB, LATEX, NumPy, Matplotlib, Pandas, Git, Linux and Windows CLI + +Coursework: Data Structures and Algorithms, Design and Analysis of Algorithms, Systems and Networks, Operating System Design, Machine Learning, Processor Design, Advanced Computer Organization \ No newline at end of file diff --git a/TalentLLM-main/resume_mmds/pyudhistira02@gmail.com.pdf.mmd b/TalentLLM-main/resume_mmds/pyudhistira02@gmail.com.pdf.mmd new file mode 100644 index 0000000000000000000000000000000000000000..629ee4d7982ff96a6afc3483c62d88d58424c9dd --- /dev/null +++ b/TalentLLM-main/resume_mmds/pyudhistira02@gmail.com.pdf.mmd @@ -0,0 +1,110 @@ +\title{ +Prama Yudhistira +} + +470-529-4451 | pyudhistira3@gatech.edu | linkedin.com/in/pramayudhistira | github.com/PramaYudhistira + +\section*{EDUCATION} + +Georgia Institute of Technology + +Atlanta, GA + +Bachelor of Science in Computer Science + +- Concentrations: Intelligence and Information Internetworks + +- Relevant Coursework: Design \& Analysis of Algorithms, Data Structures \& Algorithms, Database Systems, Computer Organization \& Programming, Discrete Mathematics, Applied Combinatorics, Linear Algebra + +- Organizations: Big Data Big Impact @ Georgia Tech, Grand Challenges Living Learning Community, Indonesian Students Association, Phi Sigma Kappa Fraternity, RoboJackets + +\section*{EXPERIENCE} + +Incoming Database Development Intern | Full-Time + +WebstaurantStore by Clark Associates Inc. + +- The position will begin in May 2024 + +Software Developer | Part-Time + +AI-based Discovery and Innovation VIP + +- Working with Dr. Ali Adibi and Emory University researchers to enhance medical imaging with Computer Vision +May 2024 - Aug. 2024 + +Lititz, $P A$ + +Aug. 2023 - Present + +Atlanta, GA + +May 2023 - Aug. 2023 + +Atlanta, GA + +Incident $I Q$ + +- Contributed to a scalable software solution that serves a userbase of over 10,000,000 across 1,000 school districts + +- Reduced the backlog by $10 \%$ within the first 2 weeks of joining by resolving critical bugs, quickly adapting to the software's robust stack comprising ASP.NET, AngularJS, and SQLServer + +- Increased team velocity by $25 \%$ by utilizing Azure DevOps to streamline CI/CD pipelines and enhancing new feature sets in an Agile environment + +- Leveraged Azure Data Studio for database manipulation to diagnose database issues, leading to the resolution of over 50 bugs throughout the internship + +- Created over 80 Jest Unit tests ensuring seamless migration from AngularJS to React + +Software Developer - Team Lead | Part-Time + +Aug. 2022 - Dec. 2022 + +Gaming for Electric Power Grids VIP + +Atlanta, GA + +- Led a team of 5 using Agile methodologies to rapidly develop and iterate a project in Unity in the programming and implementation team + +- Designed an accurate in-game wildfire algorithm with $85 \%$ accuracy based on expert feedback + +- Optimized scripts of game objects in C\#, ensuring a more efficient simulation, and created over 10 new features + +- Utilized player data to train a machine learning algorithm, giving engineers strategies on actions to take in real disaster scenarios + +\section*{Projects} + +SideHustleApp | C\#, SQL, Javascript, Angular, ASP.NET, Heroku, REST, Git + +June 2023 - Present + +An amalgamation of Craigslist and OfferUp tailored for the college student demographic + +- Developed a full-stack web application with ASP.NET for the RESTful API and Angular for the frontend + +- Created user authentication services to ensure secure access to the application + +- Managed a complex database schema using SQLite with a code-first approach using Entity Framework + +- Designed unit tests using xUnit and Jest to ensure quality of software + +Hurricane Cost Predictor | Python, Javascript, React, Flask, D3.js, REST, Git Aug. 2022 - May 2023 + +Big Data Big Impact Data Vis and Platform Sub-Team + +- Developed a project that predicts damage costs of a hurricane in the US + +- Created an interactive mapping feature where users plot the path of a hurricane with Google Map React library + +- Implemented data visualization capabilities using D3.js + +- Constructed a REST API using Flask to gather data from the machine learning model + +\section*{TeChnical Skills} + +Languages: Python (Advanced), Java (Advanced), C\#, C, SQL (MySql, SQLite, SQL Server), JavaScript, HTML/CSS + +Frameworks: React, Node.js, Flask, JUnit, ASP.NET, Angular, AngularJS, Jest, xUnit + +Developer Tools: Git, Docker, Microsoft Azure, Github, WSL, Unity, Heroku, Vercel, TensorFlow + +Libraries: Json.NET, FastHTTP, NumPy, JavaFX, Entity Framework \ No newline at end of file diff --git a/TalentLLM-main/resume_mmds/sasad3@gatech.edu.pdf.mmd b/TalentLLM-main/resume_mmds/sasad3@gatech.edu.pdf.mmd new file mode 100644 index 0000000000000000000000000000000000000000..a0c2cc44a1eb4c4182c4713fbfe8bf06433ea026 --- /dev/null +++ b/TalentLLM-main/resume_mmds/sasad3@gatech.edu.pdf.mmd @@ -0,0 +1,109 @@ +Bachelor's in Computer Science + +Atlanta, GA + +Courses: Data Structures \& Algos, Linear Algebra, Design \& Analysis of Algorithms, Computer Organiz \& Program, Objects $8 \mathcal{G}$ Design Certificates/Achievements: Deep Learning Specialization by DeepLearning.AI + +Software Eng. Virtual Experience by J.P. Morgan (Forage) + +Work Experience + +Faculty Honors Award Spring 2022, Dean's List Fall 2022 \& Spring 2023 + +Accelerating Materials Discovery with AI + +Undergraduate Researcher + +August 2023 - Present + +Atlanta, GA + +- Applied graph neural networks to accelerate the discovery of next-generation catalysts, batteries, and solar cells, reducing time-to-discovery by $40 \%$ compared to traditional methods. + +- Developed innovative data-driven algorithms for atomic-level inverse design with a goal of achieving chemical accuracy, resulting in a $20 \%$ improvement in model reliability through the incorporation of domain-specific physical constraints. + +- Engineered a scalable testing platform integrated with big data analytics and data visualization capabilities, setting new benchmarks for materials discovery research. + +Big Data Big Impact @ Georgia Tech + +August 2022 - May 2023 + +Frontend Developer + +Atlanta, GA + +- Developed the front end of a web application to visualize data collected from neural network-based damage cost estimation of US hurricanes + +- Created UI from scratch utilizing Figma and React, integrating Google API for a geo-mapping service to plot paths \& track the progress of hurricane patterns + +- Developed interactive data visualizations using React and D3.js to display complex data in an easily understandable format, resulting in a boost in user engagement with the application + +- Worked closely with the team to integrate the neural network model for estimating damage costs into the app, resulting in a $40 \%$ decrease in damage cost estimation time. + +Georgia Tech Off-Road - Baja SAE + +Jan 2022 - May 2022 + +Data Acquisition Engineer + +Atlanta, GA + +- Designed \& maintained data acquisition subsystem for efficient collection, storage \& manipulation of test data, critical to team's R\&D efforts and successfully resulting in yearly design improvements of up to $5 \%$ + +- Developed real-time system for the integration of collected and processed data yielding a $20 \%$ improvement in testing cycle times, delivering reliable insights into the car's design + +- Enhanced continuous quality assurance framework that enabled daily validation/testing process \& improved accuracy by $70 \%$, paving the way for timely product releases + +TransData + +Web Developer + +Aug 2021 - Dec 2021 + +- Created four engaging web applications for businesses, increasing user engagement and efficiency by $20 \%$ + +- Designed and developed UI/UX features with client feedback in mind, resulting in a $75 \%$ rise in customer ratings + +- Utilized innovative strategies to increase application scalability and guarantee data security through Penetration Testing + +Project Experience + +Full Stack Threads App | Next.js, Clerk, Tailwind CSS, MongoDB, Shadcn, Figma, Zod + +Jun 2023 - Present + +- Designed an intuitive UI using Figma, Next.js, and Tailwind CSS, enhanced with custom shades via Shadcn. + +- Engineered a secure user authentication and profile management system leveraging Clerk for robust backend functionality. + +- Utilized MongoDB for data storage, accommodating complex schemas and efficient data population. + +- Integrated UploadThing for file management and employed Zod for robust data validation, enhancing overall user experience. + +Face Recognition System | TensorFlow + +May 2023 - Jun 2023 + +- Implemented a high-performance ConvNet-based face verification and recognition system, utilizing FaceNet's state-of-the-art one-shot learning and triplet loss algorithms to compute 128-dimensional face encodings. + +- Tuned model parameters to enhance prediction accuracy, achieving $98 \%$ precision and recall. + +- Developed a verification mechanism that leverages L2 distance calculations to determine identity authenticity, ensuring secure access for verified individuals with an impressive accuracy threshold of 0.7. + +- Applied Neural Style Transfer techniques to augment the capabilities of the face recognition system, providing not only identity verification but also real-time artistic rendering of detected faces. + +\title{ +Leadership Experience +} + +Effective Altruism at Georgia Tech + +Organizer + +August 2023 - Present + +Atlanta, GA + +- Completed AI Safety Fundamentals Fellowship, culminating in a capstone, new approaches to Myopic Decision Theory + +- Plan to develop updates to the AI Safety Fundamentals Fellowship, along with developing new materials to raise interest for it at Georgia Tech \ No newline at end of file diff --git a/TalentLLM-main/resume_mmds/smufti3@gatech.edu.pdf.mmd b/TalentLLM-main/resume_mmds/smufti3@gatech.edu.pdf.mmd new file mode 100644 index 0000000000000000000000000000000000000000..2757aa0712b6d028b9c3aec37258075f0994ca0d --- /dev/null +++ b/TalentLLM-main/resume_mmds/smufti3@gatech.edu.pdf.mmd @@ -0,0 +1,88 @@ +\title{ +Saad Mufti +} + +Alpharetta, GA 30005 | (508)-361-8811 | smufti3@gatech.edu | U.S. Citizen + +linkedin.com/in/saad-mufti-662b2918b | github.com/Saad-Mufti | stackoverflow.com/users/13351293 + +\section*{Objective} + +Computer fanatic with an appetite for hard problems. Highly adaptive, stress tolerant. Diverse background in embedded/web/cloud software development, ML applications. Interest in silicon engineering, RTL design, reinforcement learning. + +\section*{Education} + +Georgia Institute of Technology | Atlanta, GA + +Bachelor of Science in Computer Engineering, GPA 4.0 + +Worcester Polytechnic Institute | Worcester, MA + +Transfer with 90 Credit Hours, GPA 4.0 / 4.0 +Aug 2022 - Present + +Expected Graduation (After MS): May 2025 + +Aug 2021 - Jun 2022 + +\section*{Skills} + +Programming: Python, C/C++, SystemVerilog/Verilog/VHDL, Tcl, JavaScript/Node.js, MATLAB, Java, SQL, Swift, Kotlin + +Platforms: RISC-V, Intel x86, AWS (EC2, Load Balancing), GCP (Cloud Run, Firebase, Cloud Functions, App Engine) + +Hardware: Nvidia Jetson, Raspberry Pi, ARM mbed microcontroller, FPGAs, oscilloscope, logic analyzer, TI MSP430, Arduino + +Software: Vivado, Altera Quartus, PyTorch/Tensorflow, Synopsys VCS/DVE, Docker, Cadence Virtuoso, Android Studio, Git, Flask + +\section*{Experience} + +Georgia Tech | Atlanta, GA + +RISC-V Processor Design + Tapeout + +Aug 2023 - May 2024 + +- Design and verification of an SoC including RISC-V processor, UART, SPI, CORDIC modules, using $65 \mathrm{~nm}$ TSMC PDK. + +- Theory, design, verification, test of fabricated synchronous CMOS digital circuit. Using synthesis, autoplace and route (SAPR) as industry standard tools. + +Tektronix Inc. | Beaverton, OR + +Jun - Aug 2023 + +Applications Engineering Intern + +- Researched and validated a framework (using mmWave FMCW + CNNs) for federated learning on beam prediction using low power devices (Nvidia Jetson) for test + measurement. + +- Optimized training routine (in PyTorch) to accommodate resource constrained devices, enabling inspection of model performance relating to different layer types, using GPU acceleration (CUDA + TensorRT). + +- Identified possible solutions to improve model accuracy and performance, increasing model metrics by $5-10 \%$ with $20 \%$ smaller memory footprint. + +- Helped pitch solution for object tracking using Bispectral NNs (Sanborn, 2023), flexible replacement over conventional FFT. + +Yousefi Lab @ WPI | Worcester, MA + +Reinforcement Learning and Data Pipeline Researcher + +Jun-Aug 2022 + +- Researched and validated a reinforcement learning model that fit design requirements, assisted in its development using TensorFlow, producing a proof-of-concept. + +- Orchestrated development of a data pipeline using GCP tools (Pub/Sub, Dataflow, BigQuery, Vertex Al) to ingest, preprocess, and store neural data for training and running inference on a developed RL model, demonstrating scalability. + +Shoptaki Inc. I New York City, NY (Remote) + +Aug 2021 - Aug 2022 + +Fullstack Engineer (Began as SWE Intern) + +- Led full-stack (frontend + backend) development of a demo website for newcomers in data science, using ReactJS, Flask, Express, and Arango DB. + +- Implemented CD pipelines in various development workflows using GitHub Actions and GCP Cloud Run/Build, reducing errors in manual deployment to $<5 \%$ of deployments. + +- Assisted cloud migration of various AWS services to GCP with minimal impact on service or user experience. + +\section*{Relevant Coursework} + +Data Structures + Analysis of Algorithms: Implementing and evaluating time complexity of arrays, Binary Search Trees (BSTs), Linked Lists, stacks, graph algorithms, searching/sorting, Dynamic programming, NP-Completeness, Linear Programming, Cryptography. Machine Learning (WPI, Graduate Level): Markov Chains, Maximum Likelihood Estimation, Graphical Models, Gaussian Processes, Neural Networks, Reinforcement Learning, and building a deep mathematical foundation to understand them. \ No newline at end of file diff --git a/TalentLLM-main/resume_mmds/tawfiqaliu@gmail.com.pdf.mmd b/TalentLLM-main/resume_mmds/tawfiqaliu@gmail.com.pdf.mmd new file mode 100644 index 0000000000000000000000000000000000000000..4086d6d420a9782ed1da1cd50ff1e9b3f493cc67 --- /dev/null +++ b/TalentLLM-main/resume_mmds/tawfiqaliu@gmail.com.pdf.mmd @@ -0,0 +1,77 @@ +- Relevant Coursework: Data Structures and Algorithms, Design \& Analysis of Algorithms, Intro to AI, Digital Design Lab, Machine Learning, Operating Systems + +Skills, LANGUAGES, AND AwARdS + +Programming Languages: Java, HTML/CSS, JavaScript, C, VHDL, Python, Kotlin, PostgreSQL + +Frameworks: React, Angular, Flutter + +Certifications: + +Oracle Certified Associate, Java SE 8 Programmer + +Awards: + +NJIT Programming Competition(1st in GSU) + +Honors: + +MLT CP Fellow, KP Fellow Finalist + +WORK EXPERIENCE + +Systems Engineering Intern + +Boston Scientific / Valencia, CA + +May 2023 - Aug 2023 + +- Pioneered the development of a POC desktop and mobile application with Flutter, enabling real-time battery estimates for over 2,000 devices aiding patients with chronic pain and Parkinson's + +- Coordinated UI design through Figjam and Figma, garnering positive feedback from 3 marketing teams and $2 \mathrm{PMs}$ + +- Automated a script to streamline device programming, slashing therapy programming time by an impressive $93 \%$ + +Software Engineering Intern + +Jan 2023 - May 2023 + +AudioT / Atlanta, GA + +- Engineered a desktop tool, utilized by 15 team members, that reduced human error by $20 \%$ in training a poultry health identification model + +- Employed React, Amazon S3, AWS Lambda, and PostgreSQL to construct a data schema for an application handling over 50,000 records + +Software Development Intern + +May 2022 - Aug 2022 + +Amazon / Seattle, WA + +- Revitalized an internal web-based tool, elevating efficiency in tax document automation by $63 \%$, which benefitted $40+$ financial analysts + +- Spearheaded the creation and testing of AWS Lambda functions, seamlessly integrating them with Amazon API Gateway and Cloud Formation + +Software Engineering Intern (STEP) + +Google / Remote + +May 2021 - Aug 2021 + +- Overhauled a financial dashboard that facilitated over $\$ 3$ trillion in transactions, improving productivity by $30 \%$ + +- Mitigated technical debt by identifying and rectifying 17 critical bugs, resulting in a $15 \%$ increase in codebase efficiency + +LEADERSHIP + +Founding Secretary + +GSU ProgClub + +Sept 2020 - May 2021 + +- Quadrupled club membership to $400+$ through targeted events like hackjams, speaker series, and interview prep sessions + +- Launched bi-weekly technical mock interview sessions, boosting participant confidence levels by an average of $25 \%$ + +- Conducted weekly coding workshops, resulting in a $20 \%$ increase in member engagement and practical skill development \ No newline at end of file diff --git a/TalentLLM-main/resume_mmds/tawfiqfm@gmail.com.pdf.mmd b/TalentLLM-main/resume_mmds/tawfiqfm@gmail.com.pdf.mmd new file mode 100644 index 0000000000000000000000000000000000000000..9d76ad34ae4c4e365a2659dfdeb7fe2d4ee39ca8 --- /dev/null +++ b/TalentLLM-main/resume_mmds/tawfiqfm@gmail.com.pdf.mmd @@ -0,0 +1,81 @@ +\title{ +TAWFIQ MOHAMMAD +} + +\author{ +$+1(770) 652-6679 \diamond$ Atlanta, GA \\ tawfiqfm@gmail.com $\diamond$ LinkedIn $\diamond$ tawfiqmohammad.me +} + +\section*{EDUCATION} + +Georgia Institute of Technology + +Major: B.S. in Computer Science, GPA: 3.87/4.0, + +August 2020 - December 2023 + +Relevant Coursework: Computer Networks, Data Structures and Algorithms, Deep Learning, Compilers, Robotics, High-Performance Computer Architecture, Operating Systems, Advanced Algorithms, Processor Design + +TECHNICAL SKILLS AND ADDITIONAL INFORMATION + +Technical Skills: Java, JavaScript, ReactNative, Go, Python, Django, SQL, Flutter, C/C++, Assembly + +Languages: $\quad$ English (native), Arabic (fluent), Spanish (intermediate) + +\section*{EXPERIENCE} + +Google + +May 2023 - Aug 2023 + +Software Engineer Intern + +New York, NY + +- Led the development and successful launch of a user-friendly star rating smart chip for Google Sheets, now available to millions of users. + +- Utilized my expertise in Java and Javascript to discuss and formulate the design and implementation approach of the smart chip + +- Participated in quality assurance testing for the smart chip, incorporating feedback to optimize its behavior and functionality + +Google + +May 2022 - Aug 2022 + +Software Engineer Intern + +Seattle, WA + +- Conceived and implemented a robust deployment completion classification procedure involving the cleaning and processing of over two years worth of deployment data using intricate SQL queries + +- Developed deployment completion projection formulas and calculated projections using Go + +- Led the design and implementation of a dynamic and interactive visualization dashboard using HTML, CSS, and JavaScript, serving as a key tool for over 100 users monthly + +Georgia Tech Stadium-IoPT + +Aug 2021 - Jan 2023 + +Machine Learning Engineer + +Atlanta, GA + +- Collaborating with a cross-functional team of two engineers and a project manager to develop a cutting-edge football play classifier using spatio-temporal analysis techniques + +- Utilizing Python and Tensorflow for advanced data processing, modeling, and analysis, thereby creating a football play classifier with $\mathbf{8 3 \%}$ accuracy + +\section*{PROJECTS} + +\section*{SustAIn: Hackathon Project} + +- Built a social media platform to promote sustainable habits using Django, SQLite, Bootstrap, and GPT-3.5. + +- Users can create and join challenges made by GPT-3.5, track their progress on a map, and see their friends' completed challenges on the feed + +\section*{Compiler Design} + +- Designed and developed a Compiler in Python that performs static analysis and code optimization of Tiger-IR + +- Uses optimization techniques such as redundancy elimination, instruction scheduling, and register allocation + +- Performs MIPS32 code generation from Tiger-IR and control-flow analysis based on LLVM \ No newline at end of file diff --git a/TalentLLM-main/resume_pdfs/Zahmad34@gatech.edu.pdf b/TalentLLM-main/resume_pdfs/Zahmad34@gatech.edu.pdf new file mode 100644 index 0000000000000000000000000000000000000000..97c05b7417aa91cc654a593f65cab518eff63c4c Binary files /dev/null and b/TalentLLM-main/resume_pdfs/Zahmad34@gatech.edu.pdf differ diff --git a/TalentLLM-main/resume_pdfs/azishaqui@gmail.com.pdf b/TalentLLM-main/resume_pdfs/azishaqui@gmail.com.pdf new file mode 100644 index 0000000000000000000000000000000000000000..995f07b38d162b0204fcd15a3686b40f4e92198e Binary files /dev/null and b/TalentLLM-main/resume_pdfs/azishaqui@gmail.com.pdf differ diff --git a/TalentLLM-main/resume_pdfs/dan.hus120@gmail.com.pdf b/TalentLLM-main/resume_pdfs/dan.hus120@gmail.com.pdf new file mode 100644 index 0000000000000000000000000000000000000000..0f9f28cced8fdfba1f6b4a5ee39eb13d9548a5fa Binary files /dev/null and b/TalentLLM-main/resume_pdfs/dan.hus120@gmail.com.pdf differ diff --git a/TalentLLM-main/resume_pdfs/danikhan632@gmail.com.pdf b/TalentLLM-main/resume_pdfs/danikhan632@gmail.com.pdf new file mode 100644 index 0000000000000000000000000000000000000000..ac5bfeb49692cf941048c61aaa4cc773e964ab6c Binary files /dev/null and b/TalentLLM-main/resume_pdfs/danikhan632@gmail.com.pdf differ diff --git a/TalentLLM-main/resume_pdfs/himothy2110@gmail.com.pdf b/TalentLLM-main/resume_pdfs/himothy2110@gmail.com.pdf new file mode 100644 index 0000000000000000000000000000000000000000..2b49ec9e0a8d10ed265798d94077ad9aa2eda5f0 Binary files /dev/null and b/TalentLLM-main/resume_pdfs/himothy2110@gmail.com.pdf differ diff --git a/TalentLLM-main/resume_pdfs/jamesryzhkov@gmail.com.pdf b/TalentLLM-main/resume_pdfs/jamesryzhkov@gmail.com.pdf new file mode 100644 index 0000000000000000000000000000000000000000..85c5fc1c29703e7687c876f961a5cc31364ae942 Binary files /dev/null and b/TalentLLM-main/resume_pdfs/jamesryzhkov@gmail.com.pdf differ diff --git a/TalentLLM-main/resume_pdfs/mr.zeyadalsakhi@gmail.com.pdf b/TalentLLM-main/resume_pdfs/mr.zeyadalsakhi@gmail.com.pdf new file mode 100644 index 0000000000000000000000000000000000000000..a89b7e406a5fc6994271c24e979bee014318e463 Binary files /dev/null and b/TalentLLM-main/resume_pdfs/mr.zeyadalsakhi@gmail.com.pdf differ diff --git a/TalentLLM-main/resume_pdfs/nhossain34@gatech.edu.pdf b/TalentLLM-main/resume_pdfs/nhossain34@gatech.edu.pdf new file mode 100644 index 0000000000000000000000000000000000000000..250096dac9eda19ea1fdb313e53375c983ca74de Binary files /dev/null and b/TalentLLM-main/resume_pdfs/nhossain34@gatech.edu.pdf differ diff --git a/TalentLLM-main/resume_pdfs/pyudhistira02@gmail.com.pdf b/TalentLLM-main/resume_pdfs/pyudhistira02@gmail.com.pdf new file mode 100644 index 0000000000000000000000000000000000000000..a31eb34b6f961c43f5f78be6b09d41673f0aa31d Binary files /dev/null and b/TalentLLM-main/resume_pdfs/pyudhistira02@gmail.com.pdf differ diff --git a/TalentLLM-main/resume_pdfs/sasad3@gatech.edu.pdf b/TalentLLM-main/resume_pdfs/sasad3@gatech.edu.pdf new file mode 100644 index 0000000000000000000000000000000000000000..735c656cbc13e581f5ade0773c3ced67ccc40b3f Binary files /dev/null and b/TalentLLM-main/resume_pdfs/sasad3@gatech.edu.pdf differ diff --git a/TalentLLM-main/resume_pdfs/smufti3@gatech.edu.pdf b/TalentLLM-main/resume_pdfs/smufti3@gatech.edu.pdf new file mode 100644 index 0000000000000000000000000000000000000000..5370a697b77c26e1258edef62fc1937aa8f9d899 Binary files /dev/null and b/TalentLLM-main/resume_pdfs/smufti3@gatech.edu.pdf differ diff --git a/TalentLLM-main/resume_pdfs/tawfiqaliu@gmail.com.pdf b/TalentLLM-main/resume_pdfs/tawfiqaliu@gmail.com.pdf new file mode 100644 index 0000000000000000000000000000000000000000..7690cee84399c01a7345f3a687411926aace7412 Binary files /dev/null and b/TalentLLM-main/resume_pdfs/tawfiqaliu@gmail.com.pdf differ diff --git a/TalentLLM-main/resume_pdfs/tawfiqfm@gmail.com.pdf b/TalentLLM-main/resume_pdfs/tawfiqfm@gmail.com.pdf new file mode 100644 index 0000000000000000000000000000000000000000..f48a9c7bbac24d0b324387be3ffdb2477784ad1e Binary files /dev/null and b/TalentLLM-main/resume_pdfs/tawfiqfm@gmail.com.pdf differ diff --git a/TalentLLM-main/rubric.txt b/TalentLLM-main/rubric.txt new file mode 100644 index 0000000000000000000000000000000000000000..59eaa1a0fb454b8b4685b601fd97c2ae28ae0def --- /dev/null +++ b/TalentLLM-main/rubric.txt @@ -0,0 +1,15 @@ + + + NOTE this rubric is not in any particular order + + Educational Background +PLEASE ONLY CONSIDER THEIR LATEST universities NOT EARLIER ONES +Degree Level +Work Experience (Assessed from BS Graduation Date) +-Open-source Contributions, Research, and Personal Projects: +- Project Scale and Complexity: + - Large-scale, High-impact Projects + - Mid-sized Projects + - Small Projects or Academic Assignments +- Relevance to the Role + diff --git a/TalentLLM-main/rubric_saad_proposal.txt b/TalentLLM-main/rubric_saad_proposal.txt new file mode 100644 index 0000000000000000000000000000000000000000..4045a339e86196619d7a88ec349fa2cec995727c --- /dev/null +++ b/TalentLLM-main/rubric_saad_proposal.txt @@ -0,0 +1,21 @@ + + + NOTE this rubric is not in any particular order + +Educational background important, but not essential. If all else equal, choose candidate with better school credentials +Degree Level + Work Experience (Assessed from BS Graduation Date) +- Open-source Contributions, Research, and Personal Projects: + - Personal Projects + - Contributions to Open-source + - Research Publications +project Experience +- Project Scale and Complexity: + - Large-scale, High-impact Projects + - Mid-sized Projects + - Small Projects or Academic Assignments +- Relevance to the Role: + - Directly Related + - Partially Related + - Unrelated + diff --git a/TalentLLM-main/scrappy.py b/TalentLLM-main/scrappy.py new file mode 100644 index 0000000000000000000000000000000000000000..a598431d2cfb1417895a649d5c07752a9a6acda9 --- /dev/null +++ b/TalentLLM-main/scrappy.py @@ -0,0 +1,23 @@ +import scrapy +from bs4 import BeautifulSoup +from scrapy.crawler import CrawlerProcess + +class MySpider(scrapy.Spider): + name = 'Daniyal' + start_urls = ['https://daniy.al'] + + def parse(self, response): + soup = BeautifulSoup(response.text, 'html.parser') + + text = soup.get_text(separator='\n', strip=True) + filename = 'daniyal.txt' + with open(filename, 'w') as f: + f.write(text) + self.log(f'Saved file {filename}') + + +process = CrawlerProcess() +process.crawl(MySpider) + + +process.start() diff --git a/TalentLLM-main/test.py b/TalentLLM-main/test.py new file mode 100644 index 0000000000000000000000000000000000000000..75983d464820ad8c13716db51be2add3a8b26b6f --- /dev/null +++ b/TalentLLM-main/test.py @@ -0,0 +1,81 @@ +import openai; +import json, os,sys +from dotenv import load_dotenv +load_dotenv() +from litellm import completion +from mathpix import extract_text +import gradio as gr + +# Mock candidate information +candidates = { + "candidate1": { + "name": "John Doe", + "info": "Experienced software developer with a focus on machine learning." + }, + "candidate2": { + "name": "Jane Smith", + "info": "Recent graduate with a strong background in data science." + }, + # Add more candidates as needed +} + +resume_location = 'resume_pdfs/1BXAuw6f1rDF05P734y_O7K8fYwgDVZvV.pdf' +resume_mmd = extract_text((resume_location)) + +def get_prompt(resume_mmd, chat_history, question): + history = '' + for user, bot in chat_history: + history += 'User:\n' + user + '\nResponse:\n' + bot + ' ' + return ('Given the resmue of a candidate, the previous chat history, and a question, answer the question as if you are the candidate. Keep the answers short and to the point.\n' + + 'Resume:\n\n' + resume_mmd + '\nEnd Resume\n' + + 'Chat History:\n\n' + history + '\nEnd Chat History\n' + + 'Question:\n\n' + question + '\nEnd Question\n') + +def chatbot_interface(candidate_name): + candidate = candidates.get(candidate_name, {}) + + if not candidate: + return "Candidate not found." + + candidate_info = candidate.get("info", "") + + def inference(message, history, model = 'gpt-3.5-turbo'): + try: + messages_litellm = [{"role": "user", "content": get_prompt(resume_mmd, history, message)}] + partial_message = "" + for chunk in completion(model=model, + messages=messages_litellm, + stream=True): + if 'content' in chunk['choices'][0]['delta']: + partial_message += chunk['choices'][0]['delta']['content'] + yield partial_message + except Exception as e: + print("Exception encountered:", str(e)) + yield f"An Error occured please 'Clear' the error and try your question again" + + return gr.ChatInterface( + inference, + chatbot=gr.Chatbot(height=400), + textbox=gr.Textbox(placeholder="Enter text here...", container=False, scale=5), + description=f""" + You are chatting with a resume.""", + title="Chat with Resume", + examples=["Introduce yourself."], + retry_btn="Retry", + undo_btn="Undo", + clear_btn="Clear", + ) + +def echo(message, history): + return message + +# Create Gradio app +iface = gr.Interface( + fn=chatbot_interface, + inputs=gr.Dropdown(choices=list(candidates.keys()), label="Select Candidate"), + outputs=gr.ChatInterface(echo), + live=True +) + +# Launch Gradio app +iface.launch() \ No newline at end of file
Georgia Institute of Technology +Bachelors of Science in Computer Science +August 2020 - December 2023
Link Link
import guidance + +# set the default language model used to execute guidance programs +guidance.llm = guidance.llms.TWGUI("http://127.0.0.1:5000") + +# define a guidance program that adapts a proverb +program = guidance("""Tweak this proverb to apply to model instructions instead. + +{{proverb}} +- {{book}} {{chapter}}:{{verse}} + +UPDATED +Where there is no guidance{{gen 'rewrite' stop="\\n-"}} +- GPT {{#select 'chapter'}}9{{or}}10{{or}}11{{/select}}:{{gen 'verse'}}""") + +# execute the program on a specific proverb +executed_program = program( + proverb="Where there is no guidance, a people falls,\nbut in an abundance of counselors there is safety.", + book="Proverbs", + chapter=11, + verse=14 +)
Innovative Plugin Development: Spearheaded the design and implementation of a cutting-edge AutoGPT plugin, seamlessly integrating the GPT-4 powered AutoGPT application with Alpaca Trading API to augment algorithmic trading strategies with advanced AI capabilities.
API Integration and Security: Expertly established secure and efficient connections to Alpaca’s Trading API, enabling robust trade execution, account management, and real-time data retrieval functionalities, while ensuring data integrity and compliance with industry best practices.
Enhanced Trade Management: Developed a comprehensive suite of tools for the automated placement, modification, and cancellation of diverse stock and ETF orders, including market, limit, and stop orders, resulting in a streamlined trading experience and improved operational efficiency.
Account and Portfolio Management: Implemented advanced features for real-time monitoring and management of user account details, portfolio positions, and transaction history, delivering a holistic view of financial assets and enhancing user decision-making.
Market Data and Risk Management: Provided traders with access to vital real-time and historical market data, including stock quotes and bar data, as well as corporate action insights, complemented by a robust paper trading environment for strategy testing and risk mitigation.
Github Page
BuzzOS is an Operating System built for the Intel/AMD x86_64 architecture using assembly and Rust. The operating system includes a Graphical User Interface (GUI) and is designed to provide a complete user experience.
The operating system includes user space and a mechanism for user-level processes to perform system calls to the kernel. This allows users to run applications and perform various tasks on the system.
BuzzOS also includes drivers for various hardware components, including the keyboard, mouse, timer, disk, and Intel PIC 8259. These drivers enable a robust input experience and ensure that the operating system can communicate effectively with various hardware components.
In addition to the core operating system functionality, BuzzOS also includes a fully functional desktop interface with games and system apps. This interface provides users with a familiar and intuitive environment for interacting with the operating system.
Overall, BuzzOS is an impressive project that demonstrates the power and flexibility of modern operating systems. By leveraging assembly and Rust, the project was able to create a complete operating system with a GUI and a range of drivers and applications. This is a significant achievement and represents a valuable contribution to the field of operating systems. Github Page
Designed and developed a tower defense game using the Flutter framework.
You can play the game here.
The COVID Vaccine Tracker is a tool for predicting the progress of COVID-19 vaccinations across US states. It uses data from vaccine databases and factors in state population to estimate when each state will reach an 80% vaccination rate. The project was created in March of 2021 but could potentially be modified for use with the Delta variant of COVID-19.
The model used in the project is based on a logarithmic curve. It provided fairly accurate predictions until the 50% vaccination mark but did not accurately predict the curve going logarithmic at that point. Despite this limitation, the tool still provides valuable insights into the progress of vaccinations across different US states.
https://github.com/danikhan632/tower_defense_game
Create-Cpp-App is a Command Line Interface (CLI) tool that provides an npm-like experience for building C++ applications. The tool is designed to streamline the process of building C++ apps by automating many of the repetitive and time-consuming tasks that developers typically face.
The tool is built to be intuitive and user-friendly, and it generates makefiles and automatically updates CMake files for a fast and efficient development experience. This allows developers to focus on writing code instead of worrying about the build process.
Create-Cpp-App also includes a range of built-in testing, address sanitization, benchmarking, and other tools for building production-ready C++ applications. These tools are designed to help developers ensure that their code is of high quality and performance.
Overall, Create-Cpp-App is an innovative tool that helps simplify the process of building C++ applications. By providing an npm-like experience, the tool makes it easy for developers to get started with building C++ apps and reduces the time and effort required to build high-quality, production-ready applications.
Clean Up Crew is a web application that serves as a platform for connecting small communities with local businesses. The application was built using Next.js, MongoDB, AWS S3, Google Maps API, and ReactJS.
The platform allows users to create and interact with posts in a given area. Users can post about community events, local businesses, and other topics related to their community. The application includes a sorting algorithm based on various factors such as location, user interaction, and other metrics to ensure that the most relevant content is displayed to users.
The project was developed by a team of programmers who participated in a programming competition. Over a period of 36 hours, the team worked on developing the application and implementing its various features. After the competition, the team was awarded 13th place out of 191 teams, which is a testament to their hard work and the effectiveness of the application they developed.
Overall, this project represents a valuable contribution to small communities looking to improve their localities and small businesses seeking new opportunities. The platform provides a means for these groups to connect and collaborate, and the sorting algorithm ensures that the most relevant content is displayed to users. By utilizing modern web technologies and APIs, the platform is able to provide a seamless and user-friendly experience for its users.
The Self-Driving Car project is a machine learning project that aims to simulate the behavior of a self-driving car using a Convolutional Neural Network (CNN) and computer vision techniques. The project involves constructing a virtual environment where a car can be driven autonomously using machine learning algorithms.
The CNN is used to determine the speed and angle of rotation of the simulated vehicle based on data obtained from a virtual camera. The camera captures images of the environment and feeds them into the CNN, which processes the data and outputs a prediction for the vehicle’s next move. The CNN is trained using a dataset of labeled images and their corresponding speed and steering angles.
To implement the CNN, the project utilizes a number of machine learning libraries, including Tensorflow, Keras, and NumPy. These libraries provide a range of tools for developing, training, and testing machine learning models, as well as tools for processing and analyzing large datasets.
The project also includes a testing environment where the performance of the self-driving car can be evaluated. This environment allows the user to adjust parameters such as the speed and complexity of the environment, and to observe how the car responds to different scenarios.
Overall, the Self-Driving Car project represents an exciting application of machine learning and computer vision techniques to the field of autonomous vehicles. By simulating the behavior of a self-driving car in a virtual environment, the project provides a safe and scalable platform for testing and developing new algorithms and techniques for autonomous driving.
The Amazon Shopping Clone is a web application built using the MERN stack (MongoDB, Express, React, and Node.js) and Stripe API. It mimics the design and user interface of the Amazon.com website, allowing users to browse and purchase products in a familiar environment.
One of the key features of the application is its login system, which allows users to create accounts and securely store their personal and payment information. This information is stored using MongoDB, a NoSQL database that provides a flexible and scalable data storage solution.
In addition to the login system, the application also utilizes the Stripe API to handle transactions in a secure and scalable manner. Stripe is a popular payment processing platform that provides a wide range of features for online businesses, including secure payment processing, subscription management, and fraud detection.
To ensure a smooth and intuitive user experience, the application implements a design language that closely mimics that of the Amazon.com website. This includes a consistent color scheme, typography, and layout, as well as familiar user interface elements such as navigation menus, search bars, and product listings.
Overall, the Amazon Shopping Clone provides a robust and scalable platform for online shopping that combines the familiarity and convenience of Amazon.com with the security and scalability of modern web technologies. Github Page
You can access the live demo of the FakeBlock Shopping project here