Topic #10: 오픈소스 LLM 씬의 라이징 스타! 'DeepSeek'을 알아보자 > 자유게시판

본문 바로가기

자유게시판

자유게시판 HOME


Topic #10: 오픈소스 LLM 씬의 라이징 스타! 'DeepSeek'을 알아보자

페이지 정보

profile_image
작성자 Devin
댓글 0건 조회 8회 작성일 25-02-01 18:54

본문

The DeepSeek v3 paper (and are out, after yesterday's mysterious launch of Loads of interesting details in right here. More evaluation results may be discovered right here. This is probably solely model particular, so future experimentation is required here. This mannequin is a high-quality-tuned 7B parameter LLM on the Intel Gaudi 2 processor from the Intel/neural-chat-7b-v3-1 on the meta-math/MetaMathQA dataset. The Intel/neural-chat-7b-v3-1 was originally advantageous-tuned from mistralai/Mistral-7B-v-0.1. 1.3b-instruct is a 1.3B parameter model initialized from deepseek-coder-1.3b-base and effective-tuned on 2B tokens of instruction information.

댓글목록

등록된 댓글이 없습니다.