Skip to content

Commit 36b64d7

Browse files
committed
added bib number to csv
1 parent 0647444 commit 36b64d7

File tree

1 file changed

+36
-24
lines changed

1 file changed

+36
-24
lines changed

getTopContainerCountByResource.py

Lines changed: 36 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@
1919
ids = requests.get(baseURL + endpoint, headers=headers).json()
2020

2121
f=csv.writer(open('topContainerCountByResource.csv', 'wb'))
22-
f.writerow(['title']+['uri']+['id_0']+['id_1']+['id_2']+['id_3']+['topContainerCount'])
22+
f.writerow(['title']+['bib']+['uri']+['id_0']+['id_1']+['id_2']+['id_3']+['topContainerCount'])
2323

2424
f2=csv.writer(open('topContainersLinks.csv', 'wb'))
2525
f2.writerow(['resourceUri']+['topContainerUri'])
@@ -32,14 +32,19 @@
3232
topContainerLinks = []
3333
uniqueTopContainers = []
3434
for id in ids:
35-
3635
resourceTopContainers = []
3736
print 'id', id, total, 'records remaining'
3837
total = total - 1
3938
endpoint = '/repositories/3/resources/'+str(id)
4039
output = requests.get(baseURL + endpoint, headers=headers).json()
4140
title = output['title'].encode('utf-8')
41+
print title
4242
uri = output['uri']
43+
try:
44+
bib = output['user_defined']['real_1']
45+
except:
46+
bib =''
47+
print bib
4348
id0 = output['id_0']
4449
try:
4550
id1 = output['id_1']
@@ -52,23 +57,27 @@
5257
try:
5358
id3 = output['id_3']
5459
except:
55-
id3=''
56-
# ###searchEndpoint = '/repositories/3/top_containers/search'
57-
# ###output = requests.get(baseURL + searchEndpoint, headers=headers).json()
60+
id3= ''
5861
page = 1
59-
payload = {'page': page, 'page_size': '3000', 'root_record': endpoint}
60-
search = requests.get(baseURL+'/search', headers=headers, params=payload).json()
62+
resultsPage = ''
63+
# while resultsPage != []:
64+
# payload = {'page': page, 'offset': offset, 'page_size': '100', 'root_record': endpoint}
65+
# print payload
66+
# search = requests.get(baseURL+'/search', headers=headers, params=payload).json()
67+
# print search
68+
69+
# resultsPage = search['results']
70+
# for result in resultsPage:
71+
# results.append(result)
6172
results = []
62-
resultsPage = search['results']
63-
for result in resultsPage:
64-
results.append(result)
6573
while resultsPage != []:
66-
page = page + 1
67-
payload = {'page': page, 'page_size': '3000', 'root_record': endpoint}
74+
print page
75+
payload = {'page': page, 'page_size': '100', 'root_record': endpoint}
6876
search = requests.get(baseURL+'/search', headers=headers, params=payload).json()
6977
resultsPage = search['results']
7078
for result in resultsPage:
7179
results.append(result)
80+
page = page + 1
7281

7382
for result in results:
7483
try:
@@ -85,23 +94,26 @@
8594
topContainers = []
8695
topContainerCount = len(resourceTopContainers)
8796
print 'top containers', topContainerCount
88-
f.writerow([title]+[uri]+[id0]+[id1]+[id2]+[id3]+[topContainerCount])
97+
f.writerow([title]+[bib]+[uri]+[id0]+[id1]+[id2]+[id3]+[topContainerCount])
8998

99+
print 'top container links'
90100
for topContainerLink in topContainerLinks:
91101
f2.writerow([topContainerLink[:topContainerLink.index('|')]]+[topContainerLink[topContainerLink.index('|')+1:]])
92102

103+
print 'unique top containers'
93104
for topContainer in uniqueTopContainers:
94-
search = requests.get(baseURL+topContainer, headers=headers).json()
95-
try:
96-
indicator = search['indicator']
97-
except:
98-
indicator = ''
99-
100-
try:
101-
barcode = search['barcode']
102-
except:
103-
barcode = ''
104-
f3.writerow([topContainer]+[indicator]+[barcode])
105+
print topContainer
106+
search = requests.get(baseURL+topContainer, headers=headers).json()
107+
try:
108+
indicator = search['indicator']
109+
except:
110+
indicator = ''
111+
112+
try:
113+
barcode = search['barcode']
114+
except:
115+
barcode = ''
116+
f3.writerow([topContainer]+[indicator]+[barcode])
105117

106118
elapsedTime = time.time() - startTime
107119
m, s = divmod(elapsedTime, 60)

0 commit comments

Comments
 (0)