Skip to content
View ZhilingYan's full-sized avatar

Block or report ZhilingYan

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Showing results

Tongyi Deep Research, the Leading Open-source Deep Research Agent

Python 17,454 1,340 Updated Dec 3, 2025

U-Bench: A Comprehensive Understanding of U-Net through 100-Variant Benchmarking

Python 144 17 Updated Nov 6, 2025

[ICML 2025] MedXpertQA: Benchmarking Expert-Level Medical Reasoning and Understanding

Python 130 7 Updated Jul 17, 2025

EasyR1: An Efficient, Scalable, Multi-Modality RL Training Framework based on veRL

Python 4,172 319 Updated Nov 27, 2025

Automated Quality Control and visual reports for Quality Assessment of structural (T1w, T2w) and functional MRI of the brain

Python 341 134 Updated Nov 24, 2025

Medical imaging processing for AI applications.

Python 2,330 256 Updated Dec 1, 2025
Python 1,014 62 Updated Nov 20, 2025

Codes for Visual Sketchpad: Sketching as a Visual Chain of Thought for Multimodal Language Models

Jupyter Notebook 269 15 Updated Aug 5, 2025

Agent-R1: Training Powerful LLM Agents with End-to-End Reinforcement Learning

Python 1,011 72 Updated Nov 25, 2025

Learning to Use Medical Tools with Multi-modal Agent

Python 213 24 Updated Feb 15, 2025

Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 500+ LLMs (Qwen3, Qwen3-MoE, Llama4, GLM4.5, InternLM3, DeepSeek-R1, ...) and 200+ MLLMs (Qwen3-VL, Qwen3-Omni, InternVL3.5, Ovis2.5, Llava, GLM4v, Ph…

Python 11,431 1,024 Updated Dec 3, 2025

This is the first paper to explore how to effectively use R1-like RL for MLLMs and introduce Vision-R1, a reasoning MLLM that leverages cold-start initialization and RL training to incentivize reas…

Python 732 20 Updated Sep 10, 2025

[ICLR'25] Official code for the paper 'MLLMs Know Where to Look: Training-free Perception of Small Visual Details with Multimodal LLMs'

Python 305 16 Updated Apr 20, 2025

Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)

Python 63,464 7,672 Updated Dec 1, 2025

OpenVLThinker: An Early Exploration to Vision-Language Reasoning via Iterative Self-Improvement

Python 122 6 Updated Jul 24, 2025
9 Updated Mar 28, 2025

[NAACL 2025] VividMed: Vision Language Model with Versatile Visual Grounding for Medicine

Python 26 3 Updated Mar 10, 2025
Jupyter Notebook 53 11 Updated Dec 11, 2024

Project Page For "Seg-Zero: Reasoning-Chain Guided Segmentation via Cognitive Reinforcement"

Python 569 26 Updated Jul 30, 2025
Python 38 Updated Nov 5, 2025

The official repository of paper named 'A Refer-and-Ground Multimodal Large Language Model for Biomedicine'

Python 32 Updated Nov 5, 2024

BiomedParse: A Foundation Model for Joint Segmentation, Detection, and Recognition of Biomedical Objects Across Nine Modalities

Python 603 86 Updated Oct 17, 2025

Official repository of 'Visual-RFT: Visual Reinforcement Fine-Tuning' & 'Visual-ARFT: Visual Agentic Reinforcement Fine-Tuning'’

Jupyter Notebook 2,272 101 Updated Oct 29, 2025

Solve Visual Understanding with Reinforced VLMs

Python 5,731 373 Updated Oct 21, 2025

Qwen3-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.

Jupyter Notebook 16,917 1,399 Updated Nov 28, 2025

[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4o. 接近GPT-4o表现的开源多模态对话模型

Python 9,564 738 Updated Sep 22, 2025

This is the repository of Quality Sentinel, a label quality evaluation model for medical image segmentation.

Python 22 Updated Dec 3, 2025

Export Segment Anything Models to ONNX

Python 1 Updated Oct 24, 2024
Next