Datasets:
				
			
			
	
			
	
		
			
	
		Tasks:
	
	
	
	
	Text Classification
	
	
	Modalities:
	
	
	
		
	
	Text
	
	
	Formats:
	
	
	
		
	
	parquet
	
	
	Sub-tasks:
	
	
	
	
	multi-class-classification
	
	
	Languages:
	
	
	
		
	
	English
	
	
	Size:
	
	
	
	
	100K - 1M
	
	
	License:
	
	
	
	
	
	
	
Commit 
							
							·
						
						09e718e
	
1
								Parent(s):
							
							dea2001
								
Support streaming (#1)
Browse files- Replace data file URL (1f89a9f30813984610ab912104f746c479655661)
- coarse_discourse.py +4 -4
    	
        coarse_discourse.py
    CHANGED
    
    | @@ -2,7 +2,6 @@ | |
| 2 |  | 
| 3 |  | 
| 4 | 
             
            import json
         | 
| 5 | 
            -
            import os
         | 
| 6 |  | 
| 7 | 
             
            import datasets
         | 
| 8 |  | 
| @@ -16,7 +15,8 @@ _CITATION = """\ | |
| 16 | 
             
            _DESCRIPTION = """\
         | 
| 17 | 
             
            dataset contains discourse annotation and relation on threads from reddit during 2016
         | 
| 18 | 
             
            """
         | 
| 19 | 
            -
             | 
|  | |
| 20 |  | 
| 21 |  | 
| 22 | 
             
            class CoarseDiscourse(datasets.GeneratorBasedBuilder):
         | 
| @@ -67,13 +67,13 @@ class CoarseDiscourse(datasets.GeneratorBasedBuilder): | |
| 67 | 
             
                    # TODO(coarse_discourse): Downloads the data and defines the splits
         | 
| 68 | 
             
                    # dl_manager is a datasets.download.DownloadManager that can be used to
         | 
| 69 | 
             
                    # download and extract URLs
         | 
| 70 | 
            -
                     | 
| 71 | 
             
                    return [
         | 
| 72 | 
             
                        datasets.SplitGenerator(
         | 
| 73 | 
             
                            name=datasets.Split.TRAIN,
         | 
| 74 | 
             
                            # These kwargs will be passed to _generate_examples
         | 
| 75 | 
             
                            gen_kwargs={
         | 
| 76 | 
            -
                                "filepath":  | 
| 77 | 
             
                            },
         | 
| 78 | 
             
                        ),
         | 
| 79 | 
             
                    ]
         | 
|  | |
| 2 |  | 
| 3 |  | 
| 4 | 
             
            import json
         | 
|  | |
| 5 |  | 
| 6 | 
             
            import datasets
         | 
| 7 |  | 
|  | |
| 15 | 
             
            _DESCRIPTION = """\
         | 
| 16 | 
             
            dataset contains discourse annotation and relation on threads from reddit during 2016
         | 
| 17 | 
             
            """
         | 
| 18 | 
            +
            # From: https://github.com/google-research-datasets/coarse-discourse
         | 
| 19 | 
            +
            _URL = "https://raw.githubusercontent.com/google-research-datasets/coarse-discourse/master/coarse_discourse_dataset.json"
         | 
| 20 |  | 
| 21 |  | 
| 22 | 
             
            class CoarseDiscourse(datasets.GeneratorBasedBuilder):
         | 
|  | |
| 67 | 
             
                    # TODO(coarse_discourse): Downloads the data and defines the splits
         | 
| 68 | 
             
                    # dl_manager is a datasets.download.DownloadManager that can be used to
         | 
| 69 | 
             
                    # download and extract URLs
         | 
| 70 | 
            +
                    data_path = dl_manager.download(_URL)
         | 
| 71 | 
             
                    return [
         | 
| 72 | 
             
                        datasets.SplitGenerator(
         | 
| 73 | 
             
                            name=datasets.Split.TRAIN,
         | 
| 74 | 
             
                            # These kwargs will be passed to _generate_examples
         | 
| 75 | 
             
                            gen_kwargs={
         | 
| 76 | 
            +
                                "filepath": data_path,
         | 
| 77 | 
             
                            },
         | 
| 78 | 
             
                        ),
         | 
| 79 | 
             
                    ]
         | 
