Skip to content

Commit b37b2ba

Browse files
committed
Minor updates.
1 parent 45fe135 commit b37b2ba

File tree

2 files changed

+33
-54
lines changed

2 files changed

+33
-54
lines changed

code/fetch_issues.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
#!/usr/bin/env python3
22
# -*- coding: utf-8 -*-
3-
#Copyright 2018 Luca Clissa, Marco Dalla, Roberto Morelli
3+
#Copyright 2018 Luca Clissa
44
#Licensed under the Apache License, Version 2.0 (the "License");
55
#you may not use this file except in compliance with the License.
66
#You may obtain a copy of the License at
@@ -13,7 +13,7 @@
1313
"""
1414
Created on Tue Oct 8 10:13:27 2019
1515
16-
@author: luca
16+
@author: Panos Paparrigopoulos
1717
"""
1818

1919
import urllib.request, json, sys

notebooks/rucio_data_access_HDFS.ipynb

+31-52
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,10 @@
55
"metadata": {},
66
"source": [
77
"<div class=\"alert alert-block alert-warning\">\n",
8-
"<b>Note:</b> To run this notebook you have to use \n",
8+
"<b>Note:</b> This is an adaptation of Panos' notebook for accessing Rucio data through HDFS. \n",
9+
" <br/> \n",
10+
" <br/>\n",
11+
" To run this notebook you have to use \n",
912
" <b>Swan</b>. To do so:\n",
1013
" <ul>\n",
1114
" <li>connect to <a href=\"https://swan004.cern.ch/\">https://swan004.cern.ch/</a> </li>\n",
@@ -34,48 +37,23 @@
3437
{
3538
"cell_type": "code",
3639
"execution_count": 1,
37-
"metadata": {},
40+
"metadata": {
41+
"ExecuteTime": {
42+
"end_time": "2019-10-10T14:05:57.701224Z",
43+
"start_time": "2019-10-10T14:05:57.690669Z"
44+
}
45+
},
3846
"outputs": [
3947
{
40-
"name": "stdout",
41-
"output_type": "stream",
42-
"text": [
43-
"CPU times: user 51.2 ms, sys: 36.3 ms, total: 87.5 ms\n",
44-
"Wall time: 23.4 s\n"
48+
"ename": "ModuleNotFoundError",
49+
"evalue": "No module named 'pyspark'",
50+
"output_type": "error",
51+
"traceback": [
52+
"\u001b[0;31m---------------------------------------------------------------------------\u001b[0m",
53+
"\u001b[0;31mModuleNotFoundError\u001b[0m Traceback (most recent call last)",
54+
"\u001b[0;32m<timed exec>\u001b[0m in \u001b[0;36m<module>\u001b[0;34m\u001b[0m\n",
55+
"\u001b[0;31mModuleNotFoundError\u001b[0m: No module named 'pyspark'"
4556
]
46-
},
47-
{
48-
"data": {
49-
"text/html": [
50-
"\n",
51-
" <div>\n",
52-
" <p><b>SparkSession - in-memory</b></p>\n",
53-
" \n",
54-
" <div>\n",
55-
" <p><b>SparkContext</b></p>\n",
56-
"\n",
57-
" <p><a href=\"http://db42188724fb:5201\">Spark UI</a></p>\n",
58-
"\n",
59-
" <dl>\n",
60-
" <dt>Version</dt>\n",
61-
" <dd><code>v2.4.3</code></dd>\n",
62-
" <dt>Master</dt>\n",
63-
" <dd><code>local[*]</code></dd>\n",
64-
" <dt>AppName</dt>\n",
65-
" <dd><code>Issues</code></dd>\n",
66-
" </dl>\n",
67-
" </div>\n",
68-
" \n",
69-
" </div>\n",
70-
" "
71-
],
72-
"text/plain": [
73-
"<pyspark.sql.session.SparkSession at 0x7fe9183c0128>"
74-
]
75-
},
76-
"execution_count": 1,
77-
"metadata": {},
78-
"output_type": "execute_result"
7957
}
8058
],
8159
"source": [
@@ -2354,21 +2332,22 @@
23542332
"language": "python",
23552333
"name": "python3"
23562334
},
2357-
"language_info": {
2358-
"codemirror_mode": {
2359-
"name": "ipython",
2360-
"version": 3
2361-
},
2362-
"file_extension": ".py",
2363-
"mimetype": "text/x-python",
2364-
"name": "python",
2365-
"nbconvert_exporter": "python",
2366-
"pygments_lexer": "ipython3",
2367-
"version": "3.6.5"
2368-
},
23692335
"sparkconnect": {
23702336
"bundled_options": [],
23712337
"list_of_options": []
2338+
},
2339+
"toc": {
2340+
"base_numbering": 1,
2341+
"nav_menu": {},
2342+
"number_sections": true,
2343+
"sideBar": true,
2344+
"skip_h1_title": false,
2345+
"title_cell": "Table of Contents",
2346+
"title_sidebar": "Contents",
2347+
"toc_cell": false,
2348+
"toc_position": {},
2349+
"toc_section_display": true,
2350+
"toc_window_display": false
23722351
}
23732352
},
23742353
"nbformat": 4,

0 commit comments

Comments
 (0)