(Translated by https://www.hiragana.jp/)
multimodal · GitHub Topics · GitHub
Skip to content
#

multimodal

Here are 817 public repositories matching this topic...

big-AGI

Generative AI suite powered by state-of-the-art models and providing advanced AI/AGI functions. It features AI personas, AGI functions, multi-model chats, text-to-image, voice, response streaming, code highlighting and execution, PDF import, presets for developers, much more. Deploy on-prem or in the cloud.

  • Updated Oct 11, 2024
  • TypeScript
Fengshenbang-LM

Fengshenbang-LM(ふうしん榜大模型もけい)IDEA研究けんきゅういん认知计算あずか自然しぜん语言研究けんきゅう中心ちゅうしんぬし导的だい模型もけい开源体系たいけいなり为中ぶんAIGC认知智能ちのうてきもと础设ほどこせ

  • Updated Aug 13, 2024
  • Python

Use PEFT or Full-parameter to finetune 350+ LLMs or 100+ MLLMs. (LLM: Qwen2.5, Llama3.2, GLM4, Internlm2.5, Yi1.5, Mistral, Baichuan2, DeepSeek, Gemma2, ...; MLLM: Qwen2-VL, Qwen2-Audio, Llama3.2-Vision, Llava, InternVL2, MiniCPM-V-2.6, GLM4v, Xcomposer2.5, Yi-VL, DeepSeek-VL, Phi3.5-Vision, ...)

  • Updated Oct 11, 2024
  • Python

Improve this page

Add a description, image, and links to the multimodal topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the multimodal topic, visit your repo's landing page and select "manage topics."

Learn more