The AAMC recently released its AI Vendor Evaluation Guide, a framework designed to help medical schools and residency programs assess AI tools for fairness, transparency, security, and responsible use. Since RankRx was built specifically for residency and fellowship application review, we wanted to share how our platform aligns with each of the six key categories outlined in the guide.
This blog serves as a direct response to the AAMC's vendor questions, demonstrating how RankRx works, how we safeguard applicant data, and how we empower programs to retain full control over their admissions process while benefiting from the speed and consistency of AI.
1. Balance Prediction and Understanding
AAMC's Concern:
AI should predict success while staying transparent and interpretable. Systems must measure qualities that actually align with an institution's definition of an effective student or resident.
Questions:
- How do you ensure the characteristics measured align with our institution's definition of an effective student or resident?
- How do you balance the complexity of your tool with the need for interpretable results?
- (Follow-up) How does your AI handle different data sources (e.g., academic, clinical, documents) in its decision-making?
- (Follow-up) Can you give an example of making your tool's output understandable to nontechnical users?
- (Follow-up) How do you incorporate our subject matter experts into the model-building and interpretation process?
- (Follow-up) What validation methods and metrics do you use to ensure large language model (LLM) outputs are accurate? Walk us through how you detect and prevent hallucinations or factual errors.
How RankRx Addresses This:
At RankRx, we don't decide what "success" looks like for your program, you do. Each institution defines its own success factors (e.g., USMLE scores, clinical rotations, prior research, volunteer service, or other custom experiences). RankRx then provides the framework for you to assign points, weights, and thresholds to those criteria. This ensures that all measurements align directly with your program's mission and values, not with a generic or external definition.
To maintain interpretability, RankRx avoids opaque "black-box" scoring. Every decision made by the system can be traced back to the inputs and weights your program defined. For nontechnical users, we generate clear explanations such as: "This applicant received X points for U.S. clinical experience and Y points for specialty-aligned research, based on the program's scoring rules."
Our system also supports multiple data sources, whether academic performance, clinical evaluations, or document-based materials like letters of recommendation and CVs. All are parsed, structured, and presented in a uniform, customizable scoring model.
Importantly, subject matter experts are involved from end to end: faculty and administrators define the metrics, review scoring outputs, and refine the system for each application cycle. RankRx acts as the engine that applies these definitions consistently, at scale.
Finally, to ensure reliability, we address the risk of LLM inaccuracies by grounding every output in program-defined categories and point systems. In addition to providing summary of text information, our system extracts and displays information exactly as it appears in the application. For example, with letters of recommendation, ERAS experiences, and personal statements, we provide the exact quotes from the documents so faculty can easily verify accuracy. For publications, the platform outputs the full list of citations instead of summaries, ensuring nothing is fabricated. Because all scoring is transparent and based on criteria defined by the program, any discrepancies are straightforward to identify. In addition, we run multiple assessments each cycle to validate reliability. This structured, evidence-based approach minimizes hallucinations and ensures all outputs remain directly traceable to applicant documents.
In short, RankRx makes prediction transparent and understandable because the logic is your own, our platform simply executes it with speed, consistency, and clarity.
2. Protect Against Algorithmic Bias
AAMC's Concern:
The AAMC emphasizes that AI tools in admissions must actively identify, measure, and prevent bias. Institutions want to know how vendors address fairness, representative data, communication, and accessibility.
Questions:
- What are the historical biases found in your selection tool (e.g., how are they defined and measured)? How do you prevent biases from affecting your AI tool?
- How does the AI tool ensure fairness for all demographic groups, including underrepresented in medicine?
- (Follow-up) How do you ensure your training data are representative?
- (Follow-up) How do you communicate your bias mitigation efforts to users and applicants?
- (Follow-up) How does your AI system accommodate user needs, including accessibility features and assistive technology compatibility?
How RankRx Addresses This:
RankRx is intentionally designed to minimize historical biases by removing factors that often introduce inequity in applicant screening. Our system does not rely on external "training data" or historical admissions outcomes, which can reflect past bias, but instead applies program-defined scoring criteria consistently across all applicants. Because institutions set their own weights and categories (e.g., USMLE scores, U.S. clinical experience, research), RankRx avoids embedding hidden assumptions from prior datasets.
To ensure fairness across demographic groups, applications are anonymized before processing, removing names and date of birth that could contribute to bias. Scoring is then applied uniformly, so every applicant is evaluated only on the factors programs have chosen. For transparency, programs can always trace back how points were awarded, and they can adjust scoring rules if any unintentional disparities are observed.
We communicate these fairness safeguards directly to our institutional partners, highlighting how anonymization, program-defined criteria, and transparent scoring mitigate bias risks. If programs choose, they can disclose these protections to applicants as well, reinforcing trust in the fairness of the process.
In terms of accessibility, RankRx is built to work with assistive technologies and complies with modern accessibility standards (e.g., WCAG 2.1). The platform is structured to ensure that coordinators, faculty, and administrators can easily review outputs regardless of device or ability.
In short: RankRx prevents bias not by hiding behind "black box" models but by ensuring program control, anonymization, transparency, and accessibility at every step.
3. Provide Notice and Explanation
AAMC's Concern:
The AAMC stresses that institutions must be transparent with applicants about how AI is used in admissions or residency selection. Clear disclosure helps maintain fairness, trust, and defensibility.
Key Questions:
- How do you help inform applicants about AI being used in the selection process?
- How do you advise institutions to address applicant concerns about AI being used in the admissions process?
- How well would you be able to describe the process and explain the selection tools in potential litigation?
- How would you address applicants that do not want to be screened using AI?
How RankRx Addresses This:
RankRx is built around transparency, both for institutions and applicants. We provide programs with clear documentation explaining how the platform works: scoring is based on program-defined categories and point systems, and outputs are always traceable back to the applicant's original documents.
To support institutions, RankRx offers communication templates that explain how AI is being used responsibly: anonymizing applications, applying consistent criteria, and ensuring faculty maintain full control over final decisions.
If an applicant raises concerns or does not wish to be screened using AI, RankRx is designed with flexibility — programs can review those applications outside of the platform if required.
4. Protect Data Privacy
AAMC's Concern:
The AAMC emphasizes that AI vendors must demonstrate strict safeguards for applicant data, including compliance with U.S. and international standards, minimizing data exposure, and ensuring transparency in data handling.
Key Questions:
- How do you ensure applicant data privacy and comply with U.S. guidelines (e.g., NIST Risk Management Framework) and European regulations (e.g., GDPR)?
- What processes do you recommend for allowing applicants to opt out of AI-assisted evaluation?
- How do you manage data sharing with external services or APIs, including AI tools like LLMs?
- How do you exceed compliance requirements and incorporate the latest best practices?
How RankRx Addresses This:
RankRx is built with enterprise-grade security as its foundation. Applicants' PDFs submitted by programs are only parsed for information extraction and are not stored on the RankRx platform at any time. This approach ensures minimal data exposure and reduces risk.
Our AI engine runs on OpenAI API which is SOC 2 Type 2 certified, ensuring adherence to rigorous industry standards for security and compliance. Applicant data is encrypted both in transit and at rest and is never used to train AI models. Data is retained only briefly for service integrity and abuse monitoring, and it is never made available to the public domain.
Additionally, RankRx complies with ERAS security and confidentiality requirements, allowing programs to safely integrate AI into their application review workflows without violating existing standards.
For access control, RankRx requires secure login protocols to protect sensitive applicant data. Programs control which applications they submit to the platform. If an applicant opts out of AI-assisted evaluation, the program can simply exclude that application from submission to RankRx and review it outside of the platform. This way, no applicant data is ever shared against their wishes, while programs can still use RankRx to evaluate the rest of their pool securely.
At this time, RankRx is not offered as an in-house solution. Instead, it operates on secure, SOC 2 Type 2 certified enterprise infrastructure (OpenAI API). This is something we will explore in the near future.
5. Incorporate Human Judgment
The Challenge
The AAMC emphasizes that AI should support admissions and selection, not replace human decision-making. Institutions need to see clear safeguards for oversight, training, and balance.
How RankRx Addresses This
RankRx is built to complement, not replace, human judgment. The platform applies program-defined scoring rules to ensure every application is reviewed consistently, but it does not make final admissions decisions. Instead, RankRx provides recommendations and structured outputs that help staff quickly identify applicants aligned with the program's criteria.
Programs remain fully in control: if there is ever a disagreement, the human evaluators' decision always takes precedence over AI outputs.
To ensure effective use, RankRx provides training and onboarding support for program staff and faculty. This includes walkthroughs of how scoring rules are defined, how outputs can be traced back to applicant documents, and best practices for integrating the system into the admissions workflow. Ongoing support is also available throughout each application cycle.
Subject matter experts are incorporated directly into the setup process. They determine which criteria matter, how points are assigned, and how the model should interpret application materials. This ensures the AI reflects each institution's unique mission and values.
Moreover, RankRx includes a dedicated "PD Score" column where program directors or faculty can add additional points or adjustments for individual applicants. This ensures that faculty oversight directly shapes the final score within the platform itself.
Finally, safeguards are in place to prevent over-reliance: all outputs are fully transparent and traceable to applicant documents, reminding evaluators that RankRx is a decision-support tool. Faculty and staff are encouraged to use it to streamline review, while maintaining ultimate responsibility for holistic and fair selection.
6. Monitor and Evaluate
The Challenge
The AAMC stresses that AI in admissions must be continuously monitored, audited, and improved — not just implemented once. Institutions need reassurance that vendors have processes for fairness, accountability, and alignment with evolving goals.
How RankRx Addresses This
RankRx includes ongoing monitoring and evaluation as a core feature of the platform. Each cycle, programs can review how scoring criteria performed, identify if adjustments are needed, and refine their point systems for the following year.
Because all scoring is transparent and traceable, it is straightforward to audit results for fairness and consistency, ensuring adherence to responsible AI practices.
We also gather feedback from program directors and coordinators to assess both the effectiveness of the scoring outputs and the user-friendliness of the platform. Training and support are updated regularly to reflect system improvements, so users are always aligned with the latest version without losing clarity or ease of use.
With RankRx, programs can check applicant scoring at any time during the cycle through real-time AI monitoring. If a program sees something that doesn't align with their scoring rules, they can request adjustments or fixes from us immediately. This ensures transparency and accuracy while the process is still underway.
In addition, we are expanding to participate in fellowship applications, which run on a different timeline than residency. This gives us the opportunity to test, refine, and improve the system in between residency cycles, so programs benefit from a continuously evolving platform.
Finally, we demonstrate that improvements lead to better outcomes by measuring efficiency and consistency gains (e.g., reduced screening time, clearer applicant comparisons) and by working with programs to track alignment with their institutional goals. This cycle of monitoring, feedback, and iteration ensures RankRx evolves responsibly alongside each program's needs.
Summary: How RankRx Aligns with
AAMC's Six Principles
1. Balance Prediction and Understanding
RankRx uses program-defined scoring categories and shows exact quotes and evidence, ensuring transparent, interpretable results.
2. Protect Against Algorithmic Bias
Applications are anonymized, programs set the criteria, and scoring is applied consistently, reducing bias risks.
3. Provide Notice and Explanation
We provide programs with clear communication templates to explain how AI is used, with full faculty oversight.
4. Protect Data Privacy
Built on SOC 2 Type 2 certified infrastructure, RankRx encrypts all data, never stores applicant PDFs, and complies with ERAS rules.
5. Incorporate Human Judgment
RankRx is a decision-support tool, not a decision-maker. Programs can adjust scores directly and retain control of final decisions.
6. Monitor and Evaluate
Programs can review applicant scoring in real time and request fixes. Continuous refinement between cycles ensures improvement.
Conclusion
The AAMC's vendor evaluation guide makes it clear: AI in admissions and residency selection must be transparent, fair, secure, and always support, not replace, human judgment. At RankRx, we've built our platform with these exact principles in mind.
By empowering programs to define their own criteria, maintaining strict data privacy protections, and ensuring continuous oversight and improvement, RankRx delivers an AI solution that is both responsible and practical.
As the role of AI in medical education expands, our commitment is simple: help programs save time, maintain fairness, and make better-informed decisions, without ever compromising on security or integrity.
Experience Responsible AI in Residency Selection
Ready to see how RankRx implements the AAMC's principles for responsible AI? Discover how our platform can transform your residency selection process.
Explore RankRx