| | """ |
| | Fetch data from HuggingFace dataset undertheseanlp/UVW-2026 |
| | - Get articles with quality_score >= 5 |
| | - Segment sentences using underthesea |
| | - Get first 8000 sentences |
| | """ |
| |
|
| | import re |
| | from os.path import dirname, join |
| |
|
| | from datasets import load_dataset |
| |
|
| | from underthesea import sent_tokenize, text_normalize |
| |
|
| |
|
| | def clean_text(text): |
| | """Remove formatting and clean text.""" |
| | |
| | text = text_normalize(text) |
| | |
| | text = re.sub(r'^#+\s+', '', text, flags=re.MULTILINE) |
| | |
| | text = re.sub(r'\*+', '', text) |
| | |
| | text = re.sub(r'^-+$', '', text, flags=re.MULTILINE) |
| | |
| | text = re.sub(r'\[([^\]]+)\]\([^)]+\)', r'\1', text) |
| | |
| | text = re.sub(r'\n{2,}', '\n', text) |
| | |
| | lines = [line.strip() for line in text.split('\n')] |
| | text = '\n'.join(lines) |
| | return text |
| |
|
| |
|
| | def is_valid_sentence(sent): |
| | """Check if sentence is valid for UD annotation.""" |
| | sent = sent.strip() |
| |
|
| | if not sent: |
| | return False, sent |
| | |
| | if len(sent) < 20: |
| | return False, sent |
| | |
| | if len(sent) > 300: |
| | return False, sent |
| | |
| | if not re.search(r'[àáảãạăắằẳẵặâấầẩẫậèéẻẽẹêếềểễệìíỉĩịòóỏõọôốồổỗộơớờởỡợùúủũụưứừửữựỳýỷỹỵđ]', sent, re.IGNORECASE): |
| | return False, sent |
| | |
| | if sum(1 for c in sent if c.isupper()) > len(sent) * 0.5: |
| | return False, sent |
| | |
| | if re.search(r'(bài sơ khai|sơ khai về|cần được mở rộng|Thể loại:)', sent): |
| | return False, sent |
| | |
| | if re.match(r'^(Thể loại|Danh sách|Xem thêm|Tham khảo|Liên kết ngoài|Chú thích)', sent): |
| | return False, sent |
| | |
| | if sent.count('|') > 2: |
| | return False, sent |
| | if re.search(r'\w+=\w+', sent) and sent.count('=') > 1: |
| | return False, sent |
| | |
| | if re.search(r'\[\d+\]', sent): |
| | return False, sent |
| | if re.search(r'\[cần', sent): |
| | return False, sent |
| | |
| | if re.search(r'(http|www\.|\.com|\.org)', sent, re.IGNORECASE): |
| | return False, sent |
| | |
| | num_digits = sum(1 for c in sent if c.isdigit()) |
| | if num_digits > len(sent) * 0.3: |
| | return False, sent |
| | |
| | if re.match(r'^[\*\-•]\s', sent): |
| | return False, sent |
| | return True, sent |
| |
|
| |
|
| | TARGET_COUNT = 8000 |
| |
|
| |
|
| | def fetch_and_process(): |
| | |
| | print("Loading UVW-2026 dataset from HuggingFace...") |
| | ds = load_dataset("undertheseanlp/UVW-2026", split="train") |
| |
|
| | print(f"Total articles in dataset: {len(ds)}") |
| |
|
| | |
| | print("Filtering articles by quality_score >= 5...") |
| | high_quality = [doc for doc in ds if (doc.get("quality_score") or 0) >= 5] |
| | print(f"High-quality articles: {len(high_quality)}") |
| |
|
| | |
| | print("Segmenting sentences...") |
| | all_sentences = [] |
| | for idx, doc in enumerate(high_quality): |
| | content = doc["content"] |
| | content = clean_text(content) |
| | sentences = sent_tokenize(content) |
| | for sent in sentences: |
| | sent = sent.strip() |
| | is_valid, cleaned_sent = is_valid_sentence(sent) |
| | if is_valid: |
| | all_sentences.append(cleaned_sent) |
| | if len(all_sentences) >= TARGET_COUNT: |
| | print(f"Processed {idx + 1} articles") |
| | break |
| |
|
| | |
| | sentences_out = all_sentences[:TARGET_COUNT] |
| | print(f"Total sentences collected: {len(sentences_out)}") |
| |
|
| | |
| | output_dir = dirname(dirname(__file__)) |
| | output_file = join(output_dir, "sentences_uvw.txt") |
| |
|
| | with open(output_file, "w", encoding="utf-8") as f: |
| | for i, sent in enumerate(sentences_out, 1): |
| | f.write(f"{i}\t{sent}\n") |
| |
|
| | print(f"Saved to: {output_file}") |
| |
|
| | |
| | print("\nSample sentences:") |
| | for i, sent in enumerate(sentences_out[:5], 1): |
| | print(f" {i}. {sent[:80]}...") |
| |
|
| |
|
| | if __name__ == "__main__": |
| | fetch_and_process() |
| |
|