You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

KLUE_ROBERTA_LARGE_MRC_KOR

<ν•™μŠ΅ 데이터 ꡬ좕>

κΈ°μ‘΄ KorQuAD 2.0 ν•™μŠ΅λ°μ΄ν„°λ₯Ό λ©”μΈμœΌλ‘œ ν™œμš©ν•˜μ˜€μœΌλ‚˜ Question ν˜•νƒœκ°€ μ‹€μ œ μ‚¬λžŒλ“€μ΄ 챗봇에 μž…λ ₯ν•˜λŠ” 질의 ν˜•νƒœκ°€ μ•„λ‹Œμ μ„ ν•΄κ²°ν•˜κ³ μž, 질의 ν˜•νƒœμ— λŒ€ν•΄ CHATGPTλ₯Ό ν™œμš©ν•˜μ—¬ λ‹€μ–‘ν™”ν•˜λŠ” μž‘μ—…μ„ ν•˜μ˜€μŠ΅λ‹ˆλ‹€. λΏλ§Œμ•„λ‹ˆλΌ λ™μΌν•œ 닡변을 얻을 수 μžˆλŠ” λ‹€μ–‘ν•œ ν˜•νƒœμ˜ μ§ˆλ¬Έμ„ 톡해 λͺ¨λΈμ˜ μ„±λŠ₯ ν–₯상을 μ‹œλ„ν•˜μ˜€μŠ΅λ‹ˆλ‹€. μΆ”κ°€μ μœΌλ‘œ AIHUB에 μ‘΄μž¬ν•˜λŠ” μ—¬λŸ¬ 기계독해 데이터λ₯Ό μƒ˜ν”Œλ§ν•˜μ—¬ ν•¨κ»˜ ν•™μŠ΅μ— ν™œμš©ν•˜μ˜€μŠ΅λ‹ˆλ‹€.


###ν•™μŠ΅ 데이터 νŒŒμΌμ€ λΉ„κ³΅κ°œμž…λ‹ˆλ‹€.


<λͺ¨λΈ μ„±λŠ₯>

model_name KorQuAD 2.0 KorQuAD 1.0 KLUE-MRC AIHUB(법λ₯ κΈˆμœ΅)
em f1 em f1 em f1 em f1
자체 λͺ¨λΈ(klue_roberta_large_mrc_kor) 33.55 64.43 30.5 84.7 7.6 21 29.9 87.9
timpal0l/mdeberta-v3-base-squad2 15.89 47.68 19 78.3 3.7 15.97 10.5 60.2
Dongjae/mrc2reader 17.78 55.59 20.59 83.7 4.82 19 15.4 81.3
bespin-global/klue-bert-base-aihub-mrc 1.97 12.05 12.8 49.08 3.7 24.6
ainize/klue-bert-base-mrc 6.02 19.4 26.2 75.81


λͺ¨λΈ μ„±λŠ₯ 평가λ₯Ό μœ„ν•΄ KorQuAD 2.0, KorQuAD 1.0, KLUE 벀치마크 MRC, AIHUB 법λ₯  금육 λΆ„μ•Ό 기계독해 λ“± 총 4κ°€μ§€μ˜ 평가셋을 ν™œμš©ν•˜μ—¬ ν‰κ°€ν•˜μ˜€μŠ΅λ‹ˆλ‹€.
ν˜„μž¬ ν—ˆκΉ…νŽ˜μ΄μŠ€ λͺ¨λΈ ν•œκ΅­μ–΄ 기계독해 λͺ¨λΈ 쀑 λ‹€μš΄λ‘œλ“œ μˆ˜κ°€ κ°€μž₯ λ§Žμ€ timpal0l/mdeberta-v3-base-squad2 κ³Ό 비ꡐ해도 더 λ‚˜μ€ μ„±λŠ₯을 ν™•μΈν•˜μ˜€μŠ΅λ‹ˆλ‹€.

<ν•™μŠ΅> λͺ¨λΈ ν•™μŠ΅μ€ A6000 * 2 λͺ¨λΈμ—μ„œ λŒ€λž΅ 48μ‹œκ°„ 이상 μ†Œμš”λ˜μ—ˆμŠ΅λ‹ˆλ‹€. ν•΄λ‹Ή λͺ¨λΈμ€ μžμ—°μ–΄μ²˜λ¦¬ μ „λ¬Έ 기업인 μ£Όμ‹νšŒμ‚¬ μ•„μΌλ¦¬μŠ€ν”„λŸ°ν‹°μ–΄μ˜ μ»΄ν“¨νŒ… μžμ›μ„ ν™œμš©ν•˜μ—¬ ν•™μŠ΅μ—ˆμŠ΅λ‹ˆλ‹€.

λͺ¨λΈ μ‚¬μš© μ˜ˆμ‹œ (Inference Example)

μ•„λž˜ μ½”λ“œλŠ” klue_roberta_large_mrc_kor λͺ¨λΈμ„ ν™œμš©ν•˜μ—¬ μ§ˆλ¬Έμ— λŒ€ν•œ 닡변을 μΆ”λ‘ ν•˜λŠ” μ˜ˆμ‹œμž…λ‹ˆλ‹€.
transformers 라이브러리의 pipeline을 ν™œμš©ν•©λ‹ˆλ‹€.

from transformers import pipeline

model_name = "cocoirun/klue_roberta_large_mrc_kor"

# λͺ¨λΈκ³Ό ν† ν¬λ‚˜μ΄μ € λ‘œλ“œ
qa_pipeline = pipeline(
    "question-answering",
    model=model_name,  
    tokenizer=model_name,
  max_seq_len=512,
  doc_stride=64,
  max_answer_len=256,
)

# μ˜ˆμ‹œ 질문과 λ¬Έμ„œ
question = "μΉ΄μΉ΄μ˜€λ±…ν¬λŠ” μ–΄λ–€ μ€ν–‰μΈκ°€μš”?"
context = """
μΉ΄μΉ΄μ˜€λ±…ν¬λŠ” λŒ€ν•œλ―Όκ΅­μ˜ 인터넷 μ „λ¬Έμ€ν–‰μœΌλ‘œ, λͺ¨λ°”일 μ•± 기반의 λΉ„λŒ€λ©΄ 금육 μ„œλΉ„μŠ€λ₯Ό μ œκ³΅ν•©λ‹ˆλ‹€.
κ°„νŽΈν•œ κ³„μ’Œ κ°œμ„€, μ†‘κΈˆ, λŒ€μΆœ λ“±μ˜ μ„œλΉ„μŠ€λ₯Ό μ œκ³΅ν•˜λ©°, μ Šμ€ 측을 μ€‘μ‹¬μœΌλ‘œ 큰 인기λ₯Ό μ–»κ³  μžˆμŠ΅λ‹ˆλ‹€.
"""

# μΆ”λ‘ 
result = qa_pipeline(question=question, context=context)

# κ²°κ³Ό 좜λ ₯
print(f"Answer: {result['answer']}")


** ν•΄λ‹Ή λͺ¨λΈμ€ μƒμ—…μ μœΌλ‘œ μ‚¬μš©μ€ κΈˆμ§€λ˜λ©°, ν•„μš” μ‹œμ— μ‹ μ²­ 및 메일을 λ³΄λ‚΄μ£Όμ‹œλ©΄ 확인 ν›„ μŠΉμΈλ“œλ¦¬κ² μŠ΅λ‹ˆλ‹€.
e-mail: ih.kim@aift.kr
Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for cocoirun/klue_roberta_large_mrc_kor

Base model

klue/roberta-large
Finetuned
(74)
this model