|
171 | 171 | " untar=True, cache_dir='.',\n", |
172 | 172 | " cache_subdir='')\n", |
173 | 173 | "\n", |
174 | | - "dataset_dir = os.path.join(os.path.dirname(dataset), 'aclImdb')" |
| 174 | + "dataset_dir = os.path.join(os.path.dirname(dataset), 'aclImdb_v1')" |
175 | 175 | ] |
176 | 176 | }, |
177 | 177 | { |
|
193 | 193 | }, |
194 | 194 | "outputs": [], |
195 | 195 | "source": [ |
196 | | - "train_dir = os.path.join(dataset_dir, 'train')\n", |
| 196 | + "train_dir = os.path.join(dataset_dir, 'aclImdb', 'train')\n", |
197 | 197 | "os.listdir(train_dir)" |
198 | 198 | ] |
199 | 199 | }, |
|
214 | 214 | }, |
215 | 215 | "outputs": [], |
216 | 216 | "source": [ |
217 | | - "sample_file = os.path.join(train_dir, 'pos/1181_9.txt')\n", |
| 217 | + "sample_file = os.path.join(train_dir, 'pos', '1181_9.txt')\n", |
218 | 218 | "with open(sample_file) as f:\n", |
219 | 219 | " print(f.read())" |
220 | 220 | ] |
|
286 | 286 | "seed = 42\n", |
287 | 287 | "\n", |
288 | 288 | "raw_train_ds = tf.keras.utils.text_dataset_from_directory(\n", |
289 | | - " 'aclImdb/train',\n", |
| 289 | + " train_dir,\n", |
290 | 290 | " batch_size=batch_size,\n", |
291 | 291 | " validation_split=0.2,\n", |
292 | 292 | " subset='training',\n", |
|
366 | 366 | "outputs": [], |
367 | 367 | "source": [ |
368 | 368 | "raw_val_ds = tf.keras.utils.text_dataset_from_directory(\n", |
369 | | - " 'aclImdb/train',\n", |
| 369 | + " train_dir,\n", |
370 | 370 | " batch_size=batch_size,\n", |
371 | 371 | " validation_split=0.2,\n", |
372 | 372 | " subset='validation',\n", |
|
0 commit comments