markdown
stringlengths
0
37k
code
stringlengths
1
33.3k
path
stringlengths
8
215
repo_name
stringlengths
6
77
license
stringclasses
15 values
hash
stringlengths
32
32
'What is relationship to company'? And what are the most common relationships?
recent['relationshiptocompany'] recent['relationshiptocompany'].describe() # the most common relationship to company is founder
.ipynb_checkpoints/homework7_billionaire_shengyingzhao-checkpoint.ipynb
sz2472/foundations-homework
mit
9d5197055c0e4f9ced0a9ea948ce1093
Most common source of wealth? Male vs. female?
recent['sourceofwealth'].describe() # the most common source of wealth is real estate recent.groupby('gender')['sourceofwealth'].describe() #describe the content of a given column # the most common source of wealth for male is real estate, while for female is diversified
.ipynb_checkpoints/homework7_billionaire_shengyingzhao-checkpoint.ipynb
sz2472/foundations-homework
mit
2666b68009f28574483cc1c00127db33
Given the richest person in a country, what % of the GDP is their wealth?
recent.sort_values(by='networthusbillion', ascending=False).head(10)['gdpcurrentus'] #From the website, I learned that the GDP for USA in 2014 is $17348 billion #from the previous dataframe, I learned that the richest USA billionaire made $76 billion networth richest = 76 usa_gdp = 17348 percent = round(richest / usa_gdp * 100,2) print(percent, "% of the US GDP is his wealth.")
.ipynb_checkpoints/homework7_billionaire_shengyingzhao-checkpoint.ipynb
sz2472/foundations-homework
mit
650aba2c24f6b48b93b0f68debb158ab
Add up the wealth of all of the billionaires in a given country (or a few countries) and then compare it to the GDP of the country, or other billionaires, so like pit the US vs India
recent.groupby('countrycode')['networthusbillion'].sum().sort_values(ascending=False) # USA is $2322 billion, compared to Russian is $422 billion
.ipynb_checkpoints/homework7_billionaire_shengyingzhao-checkpoint.ipynb
sz2472/foundations-homework
mit
a3a42c7f1f66920345c4570e04f0ad98
What are the most common industries for billionaires to come from? What's the total amount of billionaire money from each industry?
recent['sourceofwealth'].describe() recent.groupby('sourceofwealth')['networthusbillion'].sum().sort_values(ascending=False) How old are billionaires? How old are billionaires self made vs. non self made? or different industries? Who are the youngest billionaires? The oldest? Age distribution - maybe make a graph about it? Maybe just made a graph about how wealthy they are in general? Maybe plot their net worth vs age (scatterplot) Make a bar graph of the top 10 or 20 richest
.ipynb_checkpoints/homework7_billionaire_shengyingzhao-checkpoint.ipynb
sz2472/foundations-homework
mit
099778276aa23e285f68bdea9efecf44
How many self made billionaires vs. others?
recent['selfmade'].value_counts()
.ipynb_checkpoints/homework7_billionaire_shengyingzhao-checkpoint.ipynb
sz2472/foundations-homework
mit
c68eb58a96fda17e4da2c0c27af24477
How old are billionaires? How old are billionaires self made vs. non self made? or different industries?
recent.sort_values(by='age',ascending=False).head() columns_want = recent[['name', 'age', 'selfmade','industry']] #[[]]:dataframe columns_want.head()
.ipynb_checkpoints/homework7_billionaire_shengyingzhao-checkpoint.ipynb
sz2472/foundations-homework
mit
b3ab5d488ab05d4ddf20d455bcb70cd0
The type of variable is a tuple.
type(tuple1)
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
9400dc480de62e97205b2719af3fbcc7
Each element of a tuple can be accessed via an index. The following table represents the relationship between the index and the items in the tuple. Each element can be obtained by the name of the tuple followed by a square bracket with the index number: <img src = "https://ibm.box.com/shared/static/83kpang0opwen5e5gbwck6ktqw7btwoe.gif" width = 750, align = "center"></a> We can print out each value in the tuple:
print( tuple1[0]) print( tuple1[1]) print( tuple1[2])
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
65e17881363a2c762e395b67c1bb582d
We can print out the type of each value in the tuple:
print( type(tuple1[0])) print( type(tuple1[1])) print( type(tuple1[2]))
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
0181e80f9a048aff2c5b3011f479be61
We can also use negative indexing. We use the same table above with corresponding negative values: <img src = "https://ibm.box.com/shared/static/uwlfzo367bekwg0p5s5odxlz7vhpojyj.png" width = 750, align = "center"></a> We can obtain the last element as follows (this time we will not use the print statement to display the values):
tuple1[-1]
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
01d47810dfded4743a8663a3882582dd
We can display the next two elements as follows:
tuple1[-2] tuple1[-3]
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
30d5f14d75f6b19687d54874501e5970
We can concatenate or combine tuples by using the + sign:
tuple2=tuple1+("hard rock", 10) tuple2
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
eb215833c5a208ae708b692f878adfa9
We can slice tuples obtaining multiple values as demonstrated by the figure below: <img src = "https://ibm.box.com/shared/static/s9nofy728bcnsgnx3vh159bu16w7frnc.gif" width = 750, align = "center"></a> We can slice tuples, obtaining new tuples with the corresponding elements:
tuple2[0:3]
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
30002e6f95ebf0b2332e9caf1a169d3f
We can obtain the last two elements of the tuple:
tuple2[3:5]
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
405b629875b36d3ea84f129ff80b6530
We can obtain the length of a tuple using the length command:
len(tuple2)
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
648f0ed8bd60d522b42be496af0cce0d
This figure shows the number of elements: <img src = "https://ibm.box.com/shared/static/apxe8l3w42f597yjhizg305merlm4ijf.png" width = 750, align = "center"></a> Consider the following tuple:
Ratings =(0,9,6,5,10,8,9,6,2)
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
b829f5b668aa602854ab93d742132453
We can assign the tuple to a 2nd variable:
Ratings1=Ratings Ratings
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
fea1bdc7f150b1cdf6178e0b0304c923
We can sort the values in a tuple and save it to a new tuple:
RatingsSorted=sorted(Ratings ) RatingsSorted
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
5c7800d622d17dffe582587ad2ce795f
A tuple can contain another tuple as well as other more complex data types. This process is called 'nesting'. Consider the following tuple with several elements:
NestedT =(1, 2, ("pop", "rock") ,(3,4),("disco",(1,2)))
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
b257bc0aec798c23a1c16ec2bc27b7e7
Each element in the tuple including other tuples can be obtained via an index as shown in the figure: <img src = "https://ibm.box.com/shared/static/estqe2bczv5weocc4ag4mx9dtqy952fp.png" width = 750, align = "center"></a>
print("Element 0 of Tuple: ", NestedT[0]) print("Element 1 of Tuple: ", NestedT[1]) print("Element 2 of Tuple: ", NestedT[2]) print("Element 3 of Tuple: ", NestedT[3]) print("Element 4 of Tuple: ", NestedT[4])
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
44461dd7f44495d636227473b116efb0
We can use the second index to access other tuples as demonstrated in the figure: <img src = "https://ibm.box.com/shared/static/j1orgjuasaaj3d0feymedrnoqv8trqyo.png" width = 750, align = "center"></a> We can access the nested tuples :
print("Element 2,0 of Tuple: ", NestedT[2][0]) print("Element 2,1 of Tuple: ", NestedT[2][1]) print("Element 3,0 of Tuple: ", NestedT[3][0]) print("Element 3,1 of Tuple: ", NestedT[3][1]) print("Element 4,0 of Tuple: ", NestedT[4][0]) print("Element 4,1 of Tuple: ", NestedT[4][1])
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
825cf34238442232164866f762625662
We can access strings in the second nested tuples using a third index:
NestedT[2][1][0] NestedT[2][1][1]
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
689f47edc564f3bca700705773b711b6
We can use a tree to visualise the process. Each new index corresponds to a deeper level in the tree: <img src ='https://ibm.box.com/shared/static/vjvsygpzpwcr6czsucgno1wukyhk5vxq.gif' width = 750, align = "center"></a> Similarly, we can access elements nested deeper in the tree with a fourth index:
NestedT[4][1][0] NestedT[4][1][1]
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
e77d7aeda78331f206b94ebcfef382a7
The following figure shows the relationship of the tree and the element NestedT[4][1][1]: <img src ='https://ibm.box.com/shared/static/9y5s7515zwzc9v6i4f67yj3np2fv9evs.gif'width = 750, align = "center"></a> <a id="ref2"></a> <h2 align=center> Quiz on Tuples </h2> Consider the following tuple:
genres_tuple = ("pop", "rock", "soul", "hard rock", "soft rock", \ "R&B", "progressive rock", "disco") genres_tuple
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
31e91177ee801a46d8fd1fa3bce27739
Find the length of the tuple, "genres_tuple":
len(genres_tuple)
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
a45de868007c4ca3532a37be05fef19a
<div align="right"> <a href="#String1" class="btn btn-default" data-toggle="collapse">Click here for the solution</a> </div> <div id="String1" class="collapse"> "len(genres_tuple)" <a ><img src = "https://ibm.box.com/shared/static/n4969qbta8hhsycs2dc4n8jqbf062wdw.png" width = 1100, align = "center"></a> ``` ``` </div> Access the element, with respect to index 3:
genres_tuple[3]
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
268fc120e56866c20de345716b116ffe
<div align="right"> <a href="#2" class="btn btn-default" data-toggle="collapse">Click here for the solution</a> </div> <div id="2" class="collapse"> <a ><img src = "https://ibm.box.com/shared/static/s6r8v2uy6wifmaqv53w6adabqci47zme.png" width = 1100, align = "center"></a> </div> Use slicing to obtain indexes 3, 4 and 5:
genres_tuple[3:6]
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
4e9edf409c838947a075f7e7c61da724
<div align="right"> <a href="#3" class="btn btn-default" data-toggle="collapse">Click here for the solution</a> </div> <div id="3" class="collapse"> <a ><img src = "https://ibm.box.com/shared/static/nqo84vydw6eixdex0trybuvactcw7ffi.png" width = 1100, align = "center"></a> </div> Find the first two elements of the tuple "genres_tuple":
genres_tuple[:2]
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
c3c5a2d6caa523d508f5ace2e843b08b
<div align="right"> <a href="#q5" class="btn btn-default" data-toggle="collapse">Click here for the solution</a> </div> <div id="q5" class="collapse"> ``` genres_tuple[0:2] ``` #### Find the first index of 'disco':
genres_tuple.index("disco")
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
259233a741a4483cbaed9aa7b85b30cf
<div align="right"> <a href="#q6" class="btn btn-default" data-toggle="collapse">Click here for the solution</a> </div> <div id="q6" class="collapse"> ``` genres_tuple.index("disco") ``` <hr> #### Generate a sorted List from the Tuple C_tuple=(-5,1,-3):
C_tuple=sorted((-5, 1, -3)) C_tuple
coursera/python_for_data_science/2.1_Tuples.ipynb
mohanprasath/Course-Work
gpl-3.0
0f3b2378c6f6632061a67f04cc8e4215
另外一条规则是:位置参数优先权:
def func(a, b = 1): pass func(20, a = "G") # TypeError 对参数 a 重复赋值
Tips/2016-03-11-Arguments-and-Unpacking.ipynb
rainyear/pytips
mit
8c158b2f55b090f8a38bb179795a2c09
最保险的方法就是全部采用关键词参数。 任意参数 任意参数可以接受任意数量的参数,其中*a的形式代表任意数量的位置参数,**d代表任意数量的关键词参数:
def concat(*lst, sep = "/"): return sep.join((str(i) for i in lst)) print(concat("G", 20, "@", "Hz", sep = ""))
Tips/2016-03-11-Arguments-and-Unpacking.ipynb
rainyear/pytips
mit
12dc637098c5bcfe13dfc4468914d47b
上面的这个def concat(*lst, sep = "/")的语法是PEP 3102提出的,在 Python 3.0 之后实现。这里的关键词函数必须明确指明,不能通过位置推断:
print(concat("G", 20, "-")) # Not G-20
Tips/2016-03-11-Arguments-and-Unpacking.ipynb
rainyear/pytips
mit
1e76ecf298bec6ca4bf379de49d631c2
**d则代表任意数量的关键词参数
def dconcat(sep = ":", **dic): for k in dic.keys(): print("{}{}{}".format(k, sep, dic[k])) dconcat(hello = "world", python = "rocks", sep = "~")
Tips/2016-03-11-Arguments-and-Unpacking.ipynb
rainyear/pytips
mit
e1bc438646e793aa93a7d3ad9de42d68
Unpacking Python 3.5 添加的新特性(PEP 448),使得*a、**d可以在函数参数之外使用:
print(*range(5)) lst = [0, 1, 2, 3] print(*lst) a = *range(3), # 这里的逗号不能漏掉 print(a) d = {"hello": "world", "python": "rocks"} print({**d}["python"])
Tips/2016-03-11-Arguments-and-Unpacking.ipynb
rainyear/pytips
mit
a41d65c9bb0254157e8674f0ed09b3a9
所谓的解包(Unpacking)实际上可以看做是去掉()的元组或者是去掉{}的字典。这一语法也提供了一个更加 Pythonic 地合并字典的方法:
user = {'name': "Trey", 'website': "http://treyhunner.com"} defaults = {'name': "Anonymous User", 'page_name': "Profile Page"} print({**defaults, **user})
Tips/2016-03-11-Arguments-and-Unpacking.ipynb
rainyear/pytips
mit
34706c0f3287d8b9f4150c9ba621877b
在函数调用的时候使用这种解包的方法则是 Python 2.7 也可以使用的:
print(concat(*"ILovePython"))
Tips/2016-03-11-Arguments-and-Unpacking.ipynb
rainyear/pytips
mit
f2a59bac69fe141bcb48aca7b786ba28
Parameters The next cell initializes the parameters that are used throughout the code. They are listed as: N: The original sequence length N, which is also the length of the sequences that are going to be generated by the PFSA generated by DCGraM; drange: range of values of D for which D-Markov and DCGraM machines that will be generated; a: value up to which the autocorrelation is computed.
N = 10000000 drange = range(4,11) a = 20
dcgram.ipynb
franchenstein/dcgram
gpl-3.0
a16751e52d6b0e8100187ca64166c7d0
Original Sequence Analysis Make sure that the original sequence of length N is stored in the correct directory and run the cell to load it to X. After this, run the cells corresponding to the computation of the subsequence probabilities and the conditional probabilites for the value d_max, which is the last value in drange. Additional results can also be computed in the respective cells (autocorrelation and conditional entropy).
#Open original sequence from yaml file with open(name + '/sequences/original_len_' + str(N) + '_' + tag + '.yaml', 'r') as f: X = yaml.load(f) #Value up to which results are computed d_max = drange[-1] #Initialization of variables: p = None p_cond = None #Compute subsequence probabilities of occurrence up to length d_max p, alphabet = sa.calc_probs(X, d_max) with open(name + '/results/probabilities/original_' + tag + '.yaml', 'w') as f: yaml.dump(p, f) with open(name + '/alphabet.yaml', 'w') as f: yaml.dump(alphabet, f) #If p has been previously computed, use this cell to load the values if not p: with open(name + '/results/probabilities/original_' + tag + '.yaml', 'r') as f: p = yaml.load(f) with open(name + '/alphabet.yaml', 'r') as f: alphabet = yaml.load(f) #Compute conditional probabilities of subsequences occurring after given each symbol of the alphabet #One of the two previous cells needs to be executed first. if p: p_cond = sa.calc_cond_probs(p, alphabet, d_max - 1) with open(name + '/results/probabilities/conditional/original_' + tag + '.yaml', 'w') as f: yaml.dump(p_cond, f) else: print("Run a cell that either computes or opens the probabilities.") #If p_cond has been previously computed, use this cell to load the values if not p_cond: with open(name + '/results/probabilities/conditional/original_' + tag + '.yaml', 'r') as f: p_cond = yaml.load(f) #Compute conditional entropy if p and p_cond: h = sa.calc_cond_entropy(p, p_cond, d_max) h.to_csv(name + '/results/cond_entropies/original_' + tag + '.csv') else: print("Run the conditional probabilities cell first.") #If p_cond has been previously computed, use this cell to load the values if not h: h = pd.read_csv(name + '/results/cond_entropies/original_' + tag + '.csv') #Compute autocorrelation aut = sa.calc_autocorr(X, a) aut.to_csv(name + '/results/autocorrelations/original_' + tag + '.csv') #If aut has been previously computed, use this cell to load the values if not aut: aut = pd.read_csv(name + '/results/autocorrelations/original_' + tag + '.csv')
dcgram.ipynb
franchenstein/dcgram
gpl-3.0
107916385caebf16dee04e185494a6ce
D-Markov Machines The next step of DCGraM consists of generating D-Markov Machines for each value of D in drange defined above. The values of p_cond for each of these values is then needed, so it is necessary to compute it above. A D-Markov Machine is a PFSA with $|\Sigma|^D$ states, each one labeled with one of the subsquences of length $D$. Given a state $\omega = \sigma_1\sigma_2\ldots\sigma_D$, for each $\sigma \in \Sigma$, it transitions to the state $\sigma_2\sigma_3\ldots\sigma_D\sigma$ with probability $\Pr(\sigma|\omega)$. This is done for all states in the D-Markov machine.
dmark_machines = [] #If the D-Markov machines have not been previously created, generate them with this cell for D in list(map(str,drange)): dmark_machines.append(dmarkov.create(p_cond, D)) dmark_machines[-1].to_csv(name + '/pfsa/dmarkov_D' + D + '_' + tag + '.csv') #On the other hand, if there already are D-Markov machines, load them with this cell if not dmark_machines: for D in drange: dmark_machines.append(pd.read_csv(name + '/pfsa/dmarkov_D' + D + '_' + tag + '.csv'))
dcgram.ipynb
franchenstein/dcgram
gpl-3.0
d8772589c8b90bdafbc193be6b18539c
D-Markov Machine Analysis First of all, sequences should be generated from the D-Markov Machines. The same parameters computed in the analysis of the original sequence should be computed for the D-Markov Machines' sequences. Besides those parameters, the Kullback-Leibler Divergence and Distribution Distance between these sequences and the original sequence.
dmark_seqs = [] #Generate sequences: count = 0 for machine in dmark_machines: seq = machine.generate_sequence(N) with open(name + '/sequences/dmarkov_D' + str(drange[count]) + '_' + tag + '.yaml', 'w') as f: yaml.dump(seq, f) dmark_seqs.append(seq) count += 1 #If the sequences have been previously generated, load them here: if not dmark_seqs: for D in list(map(str,drange)): with open(name + '/sequences/dmarkov_D' + D + '_' + tag + '.yaml', 'w') as f: dmark_seqs.append(yaml.load(f)) #Compute subsequence probabilities of occurrence of the D-Markov sequences count = 0 p_dmark = [] for seq in dmark_seqs: p_dm, alphabet = sa.calc_probs(seq, d_max) p_dm.to_csv(name + '/results/probabilities/dmarkov_D'+ str(drange[count]) + '_' + tag + '.csv') p_dmark.append(p_dm) count += 1 #If p_dmark has been previously computed, use this cell to load the values if not p_dmark: for D in list(map(str,drange)): p_dm = pd.read_csv(name + '/results/probabilities/dmarkov_D' + D + '_' + tag + '.csv') p_dmark.append(p_dm) with open(name + '/alphabet.yaml', 'r') as f: alphabet = yaml.load(f) #Compute conditional probabilities of subsequences occurring after given each symbol of the alphabet #One of the two previous cells needs to be executed first. p_cond_dmark = [] count = 0 if p_dmark: for p_dm in p_dmark: p_cond_dm = sa.calc_cond_probs(p_dm, alphabet, d_max) p_cond_dm.to_csv(name + '/results/probabilities/conditional/dmarkov_D' + str(drange[count]) + '_' + tag + '.csv') p_cond_dmark.append(p_cond_dm) count += 1 else: print("Run a cell that either computes or opens the probabilities.") #If p_cond has been previously computed, use this cell to load the values if not p_cond_dmark: for D in list(map(str,drange)): p_cond_dmark.append(pd.read_csv(name + '/results/probabilities/conditional/dmarkov_D' + D + '_' + tag + '.csv')) #Compute conditional entropy count = 0 h_dmark = [] if p_dmark and p_cond_dmark: for p_dm in p_dmark: h_dm = sa.calc_cond_entropy(p_dm, p_cond_dmark[count], d_max) h_dm.to_csv(name + '/results/cond_entropies/dmarkov_D' + str(drange[count]) + '_' + tag + '.csv') h_dmark.append(h_dm) count += 1 else: print("Run the conditional probabilities cell first.") #If h_dmark has been previously computed, use this cell to load the values if not h_dmark: for D in list(map(str,drange)): h_dmark.append(pd.read_csv(name + '/results/cond_entropies/dmarkov_D' + D + '_' + tag + '.csv')) #Compute autocorrelation aut_dmark = [] count = 0 for dseq in dmark_seqs: aut_dm = sa.calc_autocorr(dseq, a) aut_dm.to_csv(name + '/results/autocorrelations/dmarkov_D' + str(drange[count]) + '_' + tag + '.csv') aut_dmark.append(aut_dm) count += 1 #If aut has been previously computed, use this cell to load the values if not aut_dmark: for D in list(map(str,drange)): aut_dmark.append(pd.read_csv(name + '/results/autocorrelations/dmarkov_D' + D + '_' + tag + '.csv')) #Compute the Kullback-Leibler Divergence between the sequences generated by the D-Markov Machines and the original #sequence. kld_dmark = [] for dseq in dmark_seqs: kld_dm = sa.calc_kld(dseq, X, d_max) kld_dmark.append(kld_dm) kld_dmark.to_csv(name + '/results/kldivergences/dmarkov_' + tag + '.csv') #If the D-Markov Kullback-Leibler divergence has been previously computed, use this cell to load the values if not kld_dmark: kld_dmark = pd.read_csv(name + '/results/kldivergences/dmarkov_' + tag + '.csv') #Compute the Probability Distances between the sequences generated by the D-Markov Machines and the original #sequence. pdist_dmark = [] for p_dm in p_dmark: pdist_dm = sa.calc_pdist(p_dm, p, d_max) pdist_dmark.append(pdist_dm) pdist_dmark.to_csv(name + '/results/prob_distances/dmarkov_' + tag + '.csv') #If the Probability Distances of the D-Markov Machines have been previously computed, load them with this cell. if not pdist_dmark: pdist_dmark = pd.read_csv(name + '/results/prob_distances/dmarkov_' + tag + '.csv')
dcgram.ipynb
franchenstein/dcgram
gpl-3.0
5b2bb722670184ce0bbe4bcfd7eb9d15
Clustering Now that we have obtained the D-Markov Machines, the next step of DCGraM is to cluster the states of these machines. For a given D-Markov Machine G$_D$, its states $q$ are considered points in a $\Sigma$-dimensional space, in which each dimension is labeled with a symbol $\sigma$ from the alphabet and the position of the state $q$ in this dimension is its probability of transitioning with this symbol. These point-states are then clustered together in $K$ clusters using a variation of the K-Means clustering algorithm that instead of using an Euclidean distance between points, uses the Kullback-Leibler Divergence between the point-state and the cluster centroids.
clustered = [] K = 4 for machine in dmark_machines: clustered.append(clustering.kmeans_kld(machine, K))
dcgram.ipynb
franchenstein/dcgram
gpl-3.0
1877cf01593685d35e7cb093fa426f8c
Graph Minimization Once that the states of the D-Markov Machines are clustered, these clusterings are then used as initial partitions of the D-Markov Machines' states. To these machines and initial partitions, a graph minimization algorithm (in the current version, only Moore) is applied in order to obtain a final reduced PFSA, the DCGraM PFSA.
dcgram_machines = [] for ini_part in clustered: dcgram_machines.append(graphmin.moore(clustered))
dcgram.ipynb
franchenstein/dcgram
gpl-3.0
753efd9e4532e1b6cab29b61c32e5139
DCGraM Analysis Now that the DCGraM machines have been generated, the same analysis done for the D-Markov Machines is used for them. Sequences are generated for each of the DCGraM machines and afterwards all of the analysis is applied to them so the comparison can be made between regular D-Markov and DCGraM.
dcgram_seqs = [] #Generate sequences: count = 0 for machine in dcgram_machines: seq = machine.generate_sequence(N) with open(name + '/sequences/dcgram_D' + str(drange[count]) + '_' + tag + '.yaml', 'w') as f: yaml.dump(seq, f) dcgram_seqs.append(seq) count += 1 #If the sequences have been previously generated, load them here: if not dcgram_seqs: for D in list(map(str,drange)): with open(name + '/sequences/dcgram_D' + D + '_' + tag + '.yaml', 'w') as f: dcgram_seqs.append(yaml.load(f)) #Compute subsequence probabilities of occurrence of the DCGraM sequences count = 0 p_dcgram = [] for seq in dcgram_seqs: p_dc, alphabet = sa.calc_probs(seq, d_max) p_dc.to_csv(name + '/results/probabilities/dcgram_D'+ str(drange[count]) + '_' + tag + '.csv') p_dcgram.append(p_dc) count += 1 #If p_dcgram has been previously computed, use this cell to load the values if not p_dcgram: for D in list(map(str,drange)): p_dc = pd.read_csv(name + '/results/probabilities/dcgram_D' + D + '_' + tag + '.csv') p_dcgram.append(p_dm) with open(name + '/alphabet.yaml', 'r') as f: alphabet = yaml.load(f) #Compute conditional probabilities of subsequences occurring after given each symbol of the alphabet #One of the two previous cells needs to be executed first. p_cond_dcgram = [] count = 0 if p_dcgram: for p_dc in p_dcgram: p_cond_dc = sa.calc_cond_probs(p_dc, alphabet, d_max) p_cond_dc.to_csv(name + '/results/probabilities/conditional/dcgram_D' + str(drange[count]) + '_' + tag + '.csv') p_cond_dcgram.append(p_cond_dc) count += 1 else: print("Run a cell that either computes or opens the probabilities.") #If p_cond_dcgram has been previously computed, use this cell to load the values if not p_cond_dcgram: for D in list(map(str,drange)): p_cond_dcgram.append(pd.read_csv(name + '/results/probabilities/conditional/dcgram_D' + D + '_' + tag + '.csv')) #Compute conditional entropy count = 0 h_dcgram = [] if p_dcgram and p_cond_dcgram: for p_dc in p_dcgram: h_dc = sa.calc_cond_entropy(p_dc, p_cond_dcgram[count], d_max) h_dc.to_csv(name + '/results/cond_entropies/dcgram_D' + str(drange[count]) + '_' + tag + '.csv') h_dcgram.append(h_dc) count += 1 else: print("Run the conditional probabilities cell first.") #If h_dcgram has been previously computed, use this cell to load the values if not h_dcgram: for D in list(map(str,drange)): h_dcgram.append(pd.read_csv(name + '/results/cond_entropies/dcgram_D' + D + '_' + tag + '.csv')) #Compute autocorrelation aut_dcgram = [] count = 0 for dcseq in dcgram_seqs: aut_dc = sa.calc_autocorr(dcseq, a) aut_dc.to_csv(name + '/results/autocorrelations/dcgram_D' + str(drange[count]) + '_' + tag + '.csv') aut_dcgram.append(aut_dc) count += 1 #If aut has been previously computed, use this cell to load the values if not aut_dcgram: for D in list(map(str,drange)): aut_dmark.append(pd.read_csv(name + '/results/autocorrelations/dcgram_D' + D + '_' + tag + '.csv')) #Compute the Kullback-Leibler Divergence between the sequences generated by the DCGraM Machines and the original #sequence. kld_dcgram = [] for dcseq in dcgram_seqs: kld_dc = sa.calc_kld(dcseq, X, d_max) kld_dcgram.append(kld_dc) kld_dcgram.to_csv(name + '/results/kldivergences/dcgram_' + tag + '.csv') #If the DCGraM Kullback-Leibler divergence has been previously computed, use this cell to load the values if not kld_dcgram: kld_dcgram = pd.read_csv(name + '/results/kldivergences/dcgram_' + tag + '.csv') #Compute the Probability Distances between the sequences generated by the DCGraM Machines and the original #sequence. pdist_dcgram = [] for p_dc in p_dcgram: pdist_dc = sa.calc_pdist(p_dc, p, d_max) pdist_dcgram.append(pdist_dc) pdist_dcgram.to_csv(name + '/results/prob_distances/dcgram_' + tag + '.csv') #If the Probability Distances of the DCGraM Machines have been previously computed, load them with this cell. if not pdist_dcgram: pdist_dcgram = pd.read_csv(name + '/results/prob_distances/dcgram_' + tag + '.csv')
dcgram.ipynb
franchenstein/dcgram
gpl-3.0
e5ec7bd5eb235aefea68aa0d9977686f
Plots Once all analysis have been made, the plots of each of those parameters is created to visualize the performance. The plots have the x-axis representing the number of states of each PFSA and the y-axis represents the parameters being observed. There are always two curves: one for the DCGraM machines and one for the D-Markov Machines. Each point in these curves represents a machine of that type for a certain value of $D$. The further right a point is in the curve, the higher its $D$-value. On the curve for conditional entropy there is also a black representing the original sequence's conditional entropy for the $L$ being used as a baseline.
#initialization import matplotlib.pyplot as plt #Labels to be used in the plots' legends labels = ['D-Markov Machines, D from ' + str(drange[0]) + ' to ' + str(d_max), 'DCGraM Machines, D from ' + str(drange[0]) + ' to ' + str(d_max), 'Original Sequence Baseline'] #Obtaining number of states of the machines to be used in the x-axis: states_dmarkov = [] for dm in dmark_machines: states_dmarkov.append(dm.shape[0]) states_dcgram = [] for dc in dcgram_machines: states_dcgram.append(dc.shape[0]) states = [states_dmarkov, states_dcgram] #Conditional Entropy plots H = 10 h_dmark_curve = [] for h_dm in h_dmarkov: h_dmark_curve.append(h_dm[H]) plt.semilogx(states[0], h_dmark_curve, marker='o', label=labels[0]) h_dcgram_curve = [] for h_dc in h_dcgram: h_dcgram_curve.append(h_dc[H]) plt.semilogx(states[1], h_dcgram_curve, marker='x', label=labels[1]) #Opening original sequence baseline: h_base = h[H] plt.axhline(y=h_base, color='k', linewidth = 3, label=labels[2]) plt.xlabel('Number of States', fontsize=16) plt.yalbel('$h_' + str(H) + '$', fontsize=16) plt.legend(loc='upper right', shadow=False, fontsize='large') plt.title('Conditional Entropy',fontsize=18,weight='bold') plt.savefig(name + '/plots/conditional_entropy_' + tag + '.eps' , bbox_inches='tight', format='eps',dpi=1000) plt.show() #Kullback-Leibler plots plt.semilogx(states[0], kld_dmark, marker='o', label=labels[0]) plt.semilogx(states[1], kld_dcgram, marker='x', label=labels[1]) plt.xlabel('Number of States', fontsize=16) plt.yalbel('$k_' + str(H) + '$', fontsize=16) plt.legend(loc='upper right', shadow=False, fontsize='large') plt.title('Kullback-Leibler Divergence',fontsize=18,weight='bold') plt.savefig(name + '/plots/kldivergence_' + tag + '.eps' , bbox_inches='tight', format='eps',dpi=1000) plt.show() #Probability Distance plots plt.semilogx(states[0], pdist_dmark, marker='o', label=labels[0]) plt.semilogx(states[1], pdist_dcgram, marker='x', label=labels[1]) plt.xlabel('Number of States', fontsize=16) plt.yalbel('$P_' + str(H) + '$', fontsize=16) plt.legend(loc='upper right', shadow=False, fontsize='large') plt.title('Probability Distance',fontsize=18,weight='bold') plt.savefig(name + '/plots/prob_distance_' + tag + '.eps' , bbox_inches='tight', format='eps',dpi=1000) plt.show() #TODO: Think how to have good plots for autocorrelation
dcgram.ipynb
franchenstein/dcgram
gpl-3.0
e0424e6bd7729ef2ac0bebe6410fd63d
Finetuning and Training
%cd $DATA_HOME_DIR #Set path to sample/ path if desired path = DATA_HOME_DIR + '/' #'/sample/' test_path = DATA_HOME_DIR + '/test/' #We use all the test data results_path=DATA_HOME_DIR + '/results/' train_path=path + '/train/' valid_path=path + '/valid/' #import Vgg16 helper class vgg = Vgg16() #Set constants. You can experiment with no_of_epochs to improve the model batch_size=64 no_of_epochs=3 #Finetune the model batches = vgg.get_batches(train_path, batch_size=batch_size) val_batches = vgg.get_batches(valid_path, batch_size=batch_size*2) vgg.finetune(batches) #Not sure if we set this for all fits vgg.model.optimizer.lr = 0.01 #Notice we are passing in the validation dataset to the fit() method #For each epoch we test our model against the validation set latest_weights_filename = None for epoch in range(no_of_epochs): print "Running epoch: %d" % epoch vgg.fit(batches, val_batches, nb_epoch=1) # latest_weights_filename = 'ft%d.h5' % epoch # vgg.model.save_weights(results_path+latest_weights_filename) print "Completed %s fit operations" % no_of_epochs
FAI_old/lesson1/dogs_cats_redux.ipynb
WNoxchi/Kaukasos
mit
fd13301ddc2cd271d3322844642da938
In the next piece of code we will cycle through our directory again: first assigning readable names to our files and storing them as a list in the variable filenames; then we will remove the case and punctuation from the text, split the words into a list of tokens, and assign the words in each file to a list in the variable corpus.
filenames = [] for files in list_textfiles('../Counting Word Frequencies/data'): files = get_filename(files) filenames.append(files) corpus = [] for filename in list_textfiles('../Counting Word Frequencies/data'): text = read_file(filename) words = text.split() clean = [w.lower() for w in words if w.isalpha()] corpus.append(clean)
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
622ce96740a06f351f7e977248640ed4
Here we recreate our list from the last exercise, counting the instances of the word privacy in each file.
for words, names in zip(corpus, filenames): print("Instances of the word \'privacy\' in", names, ":", count_in_list("privacy", words))
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
e8f55fa6f3ed0eb9f1590c376185d8e0
Next we use the len function to count the total number of words in each file.
for files, names in zip(corpus, filenames): print("There are", len(files), "words in", names)
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
f25d2678e16f5555aef9d86762890150
Now we can calculate the ratio of the word privacy to the total number of words in the file. To accomplish this we simply divide the two numbers.
print("Ratio of instances of privacy to total number of words in the corpus:") for words, names in zip(corpus, filenames): print('{:.6f}'.format(float(count_in_list("privacy", words))/(float(len(words)))),":",names)
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
f855984bc55423dd36ec6d5228b4c2f1
Now our descriptive statistics concerning word frequencies have added value. We can see that there has indeed been a steady increase in the frequency of the use of the word privacy in our corpus. When we investigate the yearly usage, we can see that the frequency almost doubled between 2008 and 2009, as well as dramatic increase between 2012 and 2014. This is also apparent in the difference between the 39th and the 40th sittings of Parliament. Let's package all of the data together so it can be displayed as a table or exported to a CSV file. First we will write our values to a list: raw contains the raw frequencies, and ratio contains the ratios. Then we will create a <span style="cursor:help;" title="a type of list where the values are permanent"><b>tuple</b></span> that contains the filename variable and includes the corresponding raw and ratio variables. Here we'll generate the ratio as a percentage.
raw = [] for i in range(len(corpus)): raw.append(count_in_list("privacy", corpus[i])) ratio = [] for i in range(len(corpus)): ratio.append('{:.3f}'.format((float(count_in_list("privacy", corpus[i]))/(float(len(corpus[i])))) * 100)) table = zip(filenames, raw, ratio)
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
2ea230e316bdedb824e0026adee6cc2d
Using the tabulate module, we will display our tuple as a table.
print(tabulate(table, headers = ["Filename", "Raw", "Ratio %"], floatfmt=".3f", numalign="left"))
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
79bd269576c3d5b91b873a55a449eed5
And finally, we will write the values to a CSV file called privacyFreqTable.
import csv with open('privacyFreqTable.csv','wb') as f: w = csv.writer(f) w.writerows(table)
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
eac6d91644b4b90ce1e0124ca96bc731
Part 2: Counting the number of transcripts Another way we can provide context is to process the corpus in a different way. Instead of splitting the data by word, we will split it in larger chunks pertaining to each individual transcript. Each transcript corresponds to a unique debate but starts with exactly the same formatting, making the files easy to split. The text below shows the beginning of a transcript. The first words are OFFICIAL REPORT (HANSARD). <img src="hansardText.png"> Here we will pass the files to another variable, called corpus_1. Instead of removing capitalization and punctuation, all we will do is split the files at every occurence of OFFICIAL REPORT (HANSARD).
corpus_1 = [] for filename in list_textfiles('../Counting Word Frequencies/data'): text = read_file(filename) words = text.split(" OFFICIAL REPORT (HANSARD)") corpus_1.append(words)
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
9f8a17d1ade5173f611ab8a13f793db5
Now, we can count the number of files in each dataset. This is also an important activity for error-checking. While it is easy to trust the numerical output of the code when it works sucessfully, we must always be sure to check that the code is actually performing in exactly the way we want it to. In this case, these numbers can be cross-referenced with the original XML data, where each transcript exists as its own file. A quick check of the directory shows that the numbers are correct.
for files, names in zip(corpus_1, filenames): print("There are", len(files), "files in", names)
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
ac564dcbd6ac9517d8e62fc6c4167b1b
Here is a screenshot of some of the raw data. We can see that there are <u>97</u> files in 2006, <u>117</u> in 2007 and <u>93</u> in 2008. The rest of the data is also correct. <img src="filecount.png"> Now we can compare the amount of occurences of privacy with the number of debates occuring in each dataset.
for names, files, words in zip(filenames, corpus_1, corpus): print("In", names, "there were", len(files), "debates. The word privacy was said", \ count_in_list('privacy', words), "times.")
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
848802cc1518f7d90576fce37324c2a5
These numbers confirm our earlier results. There is a clear indication that the usage of the term privacy is increasing, with major changes occuring between the years 2008 and 2009, as well as between 2012 and 2014. This trend is also clearly obervable between the 39th and 40th sittings of Parliament. Part 3: Looking at the corpus as a whole While chunking the corpus into pieces can help us understand the distribution or dispersion of words throughout the corpus, it's valuable to look at the corpus as a whole. Here we will create a third corpus variable corpus_3 that only contains the files named 39, 40, and 41. Note the new directory named data2. We only need these files; if we used all of the files we would literally duplicate the results.
corpus_3 = [] for filename in list_textfiles('../Counting Word Frequencies/data2'): text = read_file(filename) words = text.split() clean = [w.lower() for w in words if w.isalpha()] corpus_3.append(clean)
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
f7d30a7828b333bff51d1ffefbe520e7
Now we will combine the three lists into one large list and assign it to the variable large.
large = list(sum(corpus_3, []))
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
d141be88e2650204ba64217927a4103d
We can use the same calculations to determine the total number of occurences of privacy, as well as the total number of words in the corpus. We can also calculate the total ratio of privacy to the total number of words.
print("There are", count_in_list('privacy', large), "occurences of the word 'privacy' and a total of", \ len(large), "words.") print("The ratio of instances of privacy to total number of words in the corpus is:", \ '{:.6f}'.format(float(count_in_list("privacy", large))/(float(len(large)))), "or", \ '{:.3f}'.format((float(count_in_list("privacy", large))/(float(len(large)))) * 100),"%")
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
d2dbddd4281085f3e3f2a7fe134d398c
Another type of word frequency statistic we can generate is a type/token ratio. The types are the total number of unique words in the corpus, while the tokens are the total number of words. The type/token ratio is used to determine the variability of the language used in the text. The higher the ratio, the more complex the text will be. First we'll determine the total number of types, using <i>Python's</i> set function.
print("There are", (len(set(large))), "unique words in the Hansard corpus.")
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
eae440ca38f1264c5711623116abe669
Now we can divide the types by the tokens to determine the ratio.
print("The type/token ratio is:", ('{:.6f}'.format(len(set(large))/(float(len(large))))), "or",\ '{:.3f}'.format(len(set(large))/(float(len(large)))*100),"%")
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
cdc854c3a3872fc18a484871c6da3413
Finally, we will use the NLTK module to create a graph that shows the top 50 most frequent words in the Hansard corpus. Although privacy will not appear in the graph, it's always interesting to see what types of words are most common, and what their distribution is. NLTK will be introduced with more detail in the next section featuring concordance outputs, but here all we need to know is that we assign our variable large to the NLTK function Text in order to work with the corpus data. From there we can determine the frequency distribution for the whole text.
text = nltk.Text(large) fd = nltk.FreqDist(text)
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
eb51d2d4edbe2043b86140c9032f4995
Here we will assign the frequency distribution to the plot function to produce a graph. While it's a little hard to read, the most commonly used word in the Hansard corpus is the, with a frequency just over 400,000 occurences. The next most frequent word is to, which only has a frequency of about 225,000 occurences, almost half of the first most common word. The first 10 most frequent words appear with a much greater frequency than any of the other words in the corpus.
%matplotlib inline fd.plot(50,cumulative=False)
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
3d95099dca703f9f1e75530955c8df84
Another feature of the NLTK frequency distribution function is the generation of a list of hapaxes. These are words that appear only once in the entire corpus. While not meaningful for this study, it's an interesting way to explore the data.
fd.hapaxes()
Adding Context to Word Frequency Counts.ipynb
mediagestalt/Adding-Context
mit
4424f382cc497c1db02f7bc0754f55f7
Rating-specialized model Depending on the weights we assign, the model will encode a different balance of the tasks. Let's start with a model that only considers ratings.
# Here, configuring the model with losses and metrics. # TODO 1: Your code goes here. cached_train = train.shuffle(100_000).batch(8192).cache() cached_test = test.batch(4096).cache() # Training the ratings model. model.fit(cached_train, epochs=3) metrics = model.evaluate(cached_test, return_dict=True) print(f"Retrieval top-100 accuracy: {metrics['factorized_top_k/top_100_categorical_accuracy']:.3f}.") print(f"Ranking RMSE: {metrics['root_mean_squared_error']:.3f}.")
courses/machine_learning/deepdive2/recommendation_systems/labs/multitask.ipynb
GoogleCloudPlatform/training-data-analyst
apache-2.0
56432ed6a8a603dab02e414c9f14ae83
The model does OK on predicting ratings (with an RMSE of around 1.11), but performs poorly at predicting which movies will be watched or not: its accuracy at 100 is almost 4 times worse than a model trained solely to predict watches. Retrieval-specialized model Let's now try a model that focuses on retrieval only.
# Here, configuring the model with losses and metrics. # TODO 2: Your code goes here. # Training the retrieval model. model.fit(cached_train, epochs=3) metrics = model.evaluate(cached_test, return_dict=True) print(f"Retrieval top-100 accuracy: {metrics['factorized_top_k/top_100_categorical_accuracy']:.3f}.") print(f"Ranking RMSE: {metrics['root_mean_squared_error']:.3f}.")
courses/machine_learning/deepdive2/recommendation_systems/labs/multitask.ipynb
GoogleCloudPlatform/training-data-analyst
apache-2.0
29f728808bf9221d67c83c94fe02e001
We get the opposite result: a model that does well on retrieval, but poorly on predicting ratings. Joint model Let's now train a model that assigns positive weights to both tasks.
# Here, configuring the model with losses and metrics. # TODO 3: Your code goes here. # Training the joint model. model.fit(cached_train, epochs=3) metrics = model.evaluate(cached_test, return_dict=True) print(f"Retrieval top-100 accuracy: {metrics['factorized_top_k/top_100_categorical_accuracy']:.3f}.") print(f"Ranking RMSE: {metrics['root_mean_squared_error']:.3f}.")
courses/machine_learning/deepdive2/recommendation_systems/labs/multitask.ipynb
GoogleCloudPlatform/training-data-analyst
apache-2.0
732ca245e5823fd280e9bd36cd720c13
Proper use of Matplotlib We will use interactive plots inline in the notebook. This feature is enabled through:
%matplotlib import matplotlib.pyplot as plt import numpy as np # define a figure which can contains several plots, you can define resolution and so on here... fig2 = plt.figure() # add one axis, axes are actual plots where you can put data.fits (nx, ny, index) ax = fig2.add_subplot(1, 1, 1)
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
9b21043fd6ab12278f885110c3384ae8
Add a cruve with a title to the plot
x = np.linspace(0, 2*np.pi) ax.plot(x, np.sin(x), '+') ax.set_title('this title') plt.show() # is a simpler syntax to add one axis into the figure (we will stick to this) fig, ax = plt.subplots() ax.plot(x, np.sin(x), '+') ax.set_title('simple subplot')
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
fec543326104d7bceb4e4fa2060dbcd2
A long list of markers can be found at http://matplotlib.org/api/markers_api.html as for the colors, there is a nice discussion at http://stackoverflow.com/questions/22408237/named-colors-in-matplotlib All the components of a figure can be accessed throught the 'Figure' object
print(type(fig)) print(dir(fig)) print(fig.axes) print('This is the x-axis object', fig.axes[0].xaxis) print('And this is the y-axis object', fig.axes[0].yaxis) # arrow pointing to the origin of the axes ax_arrow = ax.annotate('ax = fig.axes[0]', xy=(0, -1), # tip of the arrow xytext=(1, -0.5), # location of the text arrowprops={'facecolor':'red', 'shrink':0.05}) # arrow pointing to the x axis x_ax_arrow = ax.annotate('ax.xaxis', xy=(3, -1), # tip of the arrow xytext=(3, -0.5), # location of the text arrowprops={'facecolor':'red', 'shrink':0.05}) xax = ax.xaxis # arrow pointing to the y axis y_ax_arrow = ax.annotate('ax.yaxis', xy=(0, 0), # tip of the arrow xytext=(1, 0.5), # location of the text arrowprops={'facecolor':'red', 'shrink':0.05})
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
809d3c67129d5233b44b791f3335405c
Add a labels to the x and y axes
# add some ascii text label # this is equivelant to: # ax.set_xlabel('x') xax.set_label_text('x') # add latex rendered text to the y axis ax.set_ylabel('$sin(x)$', size=20, color='g', rotation=0)
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
ee39b313cfc6fc5b2e56af45b1904baa
Finally dump the figure to a png file
fig.savefig('myplot.png') !ls !eog myplot.png
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
7ddd567f71ace42305755581f3ff45a5
Lets define a function that creates an empty base plot to which we will add stuff for each demonstration. The function returns the figure and the axes object.
from matplotlib import pyplot as plt import numpy as np def create_base_plot(): fig, ax = plt.subplots() ax.set_title('sample figure') return fig, ax def plot_something(): fig, ax = create_base_plot() x = np.linspace(0, 2*np.pi) ax.semilogx(x, np.cos(x)*np.cos(x/2), 'r--.') plt.show()
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
84068c53f7e9a5519391c190bffbb59f
Log plots
fig, ax = create_base_plot() # normal-xlog plots ax.semilogx(x, np.cos(x)*np.cos(x/2), 'r--.') # clear the plot and plot a function using the y axis in log scale ax.clear() ax.semilogy(x, np.exp(x)) # you can (un)set it, whenever you want #ax.set_yscale('linear') # change they y axis to linear scale #ax.set_yscale('log') # change the y axis to log scale # you can also make loglog plots #ax.clear() #ax.loglog(x, np.exp(x)*np.sin(x)) plt.setp(ax, **dict(yscale='log', xscale='log'))
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
12926edc54ca2242d8b257d709dd7799
This is equivelant to: ax.plot(x, np.exp(x)*np.sin(x)) plt.setp(ax, 'yscale', 'log', 'xscale', 'log') here we have introduced a new method of setting property values via pyplot.setp. setp takes as first argument a matplotlib object. Each pair of positional argument after that is treated as a key value pair for the set method name and its value. For example: ax.set_scale('linear') becomes plt.setp(ax, 'scale', 'linear') This is useful if you need to set lots of properties, such as:
plt.setp(ax, 'xscale', 'linear', 'xlim', [1, 5], 'ylim', [0.1, 10], 'xlabel', 'x', 'ylabel', 'y', 'title', 'foo', 'xticks', [1, 2, 3, 4, 5], 'yticks', [0.1, 1, 10], 'yticklabels', ['low', 'medium', 'high'])
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
f9fb8127d8552efc5571c4461a57aa3a
Histograms
fig1, ax = create_base_plot() n, bins, patches = ax.hist(np.random.normal(0, 0.1, 10000), bins=50)
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
f33dbdc0b89ae75c73679ac9e7c19024
Subplots Making subplots is relatively easy. Just pass the shape of the grid of plots to plt.subplots() that was used in the above examples.
# Create one figure with two plots/axes, with their xaxis shared fig, (ax1, ax2) = plt.subplots(2, sharex=True) ax1.plot(x, np.sin(x), '-.', color='r', label='first line') other = ax2.plot(x, np.cos(x)*np.cos(x/2), 'o-', linewidth=3, label='other') ax1.legend() ax2.legend() # adjust the spacing between the axes fig.subplots_adjust(hspace=0.0) # add a scatter plot to the first axis ax1.scatter(x, np.sin(x)+np.random.normal(0, 0.1, np.size(x)))
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
22cb1636fb83b94bb78020a1f55444d3
create a 3x3 grid of plots
fig, axs = plt.subplots(3, 3) print(axs.shape) # add an index to all the subplots for ax_index, ax in enumerate(axs.flatten()): ax.set_title(ax_index) # remove all ticks for ax in axs.flatten(): plt.setp(ax, 'xticks', [], 'yticks', []) fig.subplots_adjust(hspace=0, wspace=0) # plot a curve in the diagonal subplots for ax, func in zip(axs.diagonal(), [np.sin, np.cos, np.exp]): ax.plot(x, func(x))
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
f779e36a88096ec7c4c71f3718335b13
Images and contours
xx, yy = np.mgrid[-2:2:100j, -2:2:100j] img = np.sin(xx) + np.cos(yy) fig, ax = create_base_plot() # to have 0,0 in the lower left corner and no interpolation img_plot = ax.imshow(img, origin='lower', interpolation='None') # to add a grid to any axis ax.grid() img_plot.set_cmap('hot') # changing the colormap img_plot.set_cmap('spectral') # changing the colormap colorb = fig.colorbar(img_plot) # adding a color bar img_plot.set_clim(-0.5, 0.5) # changing the dynamical range # add contour levels img_contours = ax.contour(img, [-1, -0.5, 0.0, 0.5]) plt.clabel(img_contours, inline=True, fontsize=20)
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
b3d35a09a9acba74c79a49c12c07abad
Animation
from IPython.display import HTML import matplotlib.animation as animation def f(x, y): return np.sin(x) + np.cos(y) fig, ax = create_base_plot() im = ax.imshow(f(xx, yy), cmap=plt.get_cmap('viridis')) def updatefig(*args): global xx, yy xx += np.pi / 15. yy += np.pi / 20. im.set_array(f(xx, yy)) return im, ani = animation.FuncAnimation(fig, updatefig, interval=50, blit=True) _ = ani.to_html5_video() # change title during animation!! ax.set_title('runtime title')
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
8300e4ded0354d6a05f27dc9c9c4d3d6
Styles Configuring matplotlib Most of the matplotlib code chunk that are written are usually about styling and not actual plotting. One feature that might be of great help if you are in this case is to use the matplotlib.style module. In this notebook, we will go through the available matplotlib styles and their corresponding configuration files. Then we will explain the two ways of using the styles and finally show you how to write a personalized style. Pre-configured style files An available variable returns a list of the names of some pre-configured matplotlib style files.
print('\n'.join(plt.style.available)) x = np.arange(0, 10, 0.01) def f(x, t): return np.sin(x) * np.exp(1 - x / 10 + t / 2) def simple_plot(style): plt.figure() with plt.style.context(style, after_reset=True): for t in range(5): plt.plot(x, f(x, t)) plt.title('Simple plot') simple_plot('ggplot') simple_plot('dark_background') simple_plot('grayscale') simple_plot('fivethirtyeight') simple_plot('bmh')
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
416162bb368107f26309d5ac68ecf5ed
Content of the style files A matplotlib style file is a simple text file containing the desired matplotlib rcParam configuration, with the .mplstyle extension. Let's display the content of the 'ggplot' style.
import os ggplotfile = os.path.join(plt.style.core.BASE_LIBRARY_PATH, 'ggplot.mplstyle') !cat $ggplotfile
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
52c197bbdcd4fa99f995334fab7b826d
Maybe the most interesting feature of this style file is the redefinition of the color cycle using hexadecimal notation. This allows the user to define is own color palette for its multi-line plots. use versus context There are two ways of using the matplotlib styles. plt.style.use(style) plt.style.context(style): The use method applied at the beginning of a script will be the default choice in most cases when the style is to be set for the entire script. The only issue is that it sets the matplotlib style for the given Python session, meaning that a second call to use with a different style will only apply new style parameters and not reset the first style. That is if the axes.grid is set to True by the first style and there is nothing concerning the grid in the second style config, the grid will remain set to True which is not matplotlib default. On the contrary, the context method will be useful when only one or two figures are to be set to a given style. It shall be used with the with statement to create a context manager in which the plot will be made. Let's illustrate this.
plt.style.use('ggplot') plt.figure() plt.plot(x, f(x, 0))
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
24bd9cf1dfd5fb0094e0a2194adf71ff
The 'ggplot' style has been applied to the current session. One of its features that differs from standard matplotlib configuration is to put the ticks outside the main figure (axes.axisbelow: True)
with plt.style.context('dark_background'): plt.figure() plt.plot(x, f(x, 1))
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
a9525bc274bf7e1e8c37613732dbebf8
Now using the 'dark_background' style as a context, we can spot the main changes (background, line color, axis color) and we can also see the outside ticks, although they are not part of this particular style. This is the 'ggplot' axes.axisbelow setup that has not been overwritten by the new style. Once the with block has ended, the style goes back to its previous status, that is the 'ggplot' style.
plt.figure() plt.plot(x, f(x, 2))
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
e7333128436ba33e407d6403c00f660d
Custom style file Starting from these configured files, it is easy to now create our own styles for textbook figures and talk figures and switch from one to another in a single code line plt.style.use('mystyle') at the beginning of the plotting script. Where to create it ? matplotlib will look for the user style files at the following path :
print(plt.style.core.USER_LIBRARY_PATHS)
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
debf5ab945344842898013e85db7c757
Note: The directory corresponding to this path will most probably not exist so one will need to create it.
styledir = plt.style.core.USER_LIBRARY_PATHS[0] !mkdir -p $styledir
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
4eb6279fb81ddaebeb2ea620b67559ae
One can now copy an existing style file to serve as a boilerplate.
mystylefile = os.path.join(styledir, 'mystyle.mplstyle') !cp $ggplotfile $mystylefile !cd $styledir %%file mystyle.mplstyle font.size: 16.0 # large font axes.linewidth: 2 axes.grid: True axes.titlesize: x-large axes.labelsize: x-large axes.labelcolor: 555555 axes.axisbelow: True xtick.color: 555555 xtick.direction: out ytick.color: 555555 ytick.direction: out grid.color: white grid.linestyle: : # dotted line
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
f6d6ec513236e42c843a4cdb41ad1c1b
D3
%matplotlib inline import matplotlib.pyplot as plt import numpy as np import mpld3 mpld3.enable_notebook() # Scatter points fig, ax = plt.subplots(subplot_kw=dict(axisbg='#EEEEEE')) ax.grid(color='white', linestyle='solid') N = 50 scatter = ax.scatter(np.random.normal(size=N), np.random.normal(size=N), c=np.random.random(size=N), s = 1000 * np.random.random(size=N), alpha=0.3, cmap=plt.cm.jet) ax.set_title("D3 Scatter Plot", size=18); import mpld3 mpld3.display(fig) from mpld3 import plugins fig, ax = plt.subplots(subplot_kw=dict(axisbg='#EEEEEE')) ax.grid(color='white', linestyle='solid') N = 50 scatter = ax.scatter(np.random.normal(size=N), np.random.normal(size=N), c=np.random.random(size=N), s = 1000 * np.random.random(size=N), alpha=0.3, cmap=plt.cm.jet) ax.set_title("D3 Scatter Plot (with tooltips!)", size=20) labels = ['point {0}'.format(i + 1) for i in range(N)] fig.plugins = [plugins.PointLabelTooltip(scatter, labels)]
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
77af8f31a9964272b7aa05809a1846b1
Seaborn
%matplotlib plot_something() import seaborn plot_something()
notebooks/03-Plotting.ipynb
aboucaud/python-euclid2016
bsd-3-clause
1b097483dff83a06bf2d3c7fee447812
Packet Forwarding This category of questions allows you to query how different types of traffic is forwarded by the network and if endpoints are able to communicate. You can analyze these aspects in a few different ways. Traceroute Bi-directional Traceroute Reachability Bi-directional Reachability Loop detection Multipath Consistency for host-subnets Multipath Consistency for router loopbacks
bf.set_network('generate_questions') bf.set_snapshot('generate_questions')
docs/source/notebooks/forwarding.ipynb
batfish/pybatfish
apache-2.0
df2317fb9f334928fd085dcbe774ee0e
Traceroute Traces the path(s) for the specified flow. Performs a virtual traceroute in the network from a starting node. A destination IP and ingress (source) node must be specified. Other IP headers are given default values if unspecified. Unlike a real traceroute, this traceroute is directional. That is, for it to succeed, the reverse connectivity is not needed. This feature can help debug connectivity issues by decoupling the two directions. Inputs Name | Description | Type | Optional | Default Value --- | --- | --- | --- | --- startLocation | Location (node and interface combination) to start tracing from. | LocationSpec | False | headers | Packet header constraints. | HeaderConstraints | False | maxTraces | Limit the number of traces returned. | int | True | ignoreFilters | If set, filters/ACLs encountered along the path are ignored. | bool | True | Invocation
result = bf.q.traceroute(startLocation='@enter(as2border1[GigabitEthernet2/0])', headers=HeaderConstraints(dstIps='2.34.201.10', srcIps='8.8.8.8')).answer().frame()
docs/source/notebooks/forwarding.ipynb
batfish/pybatfish
apache-2.0
b2660600ab0a9336bf593bee6f6abc12
Return Value Name | Description | Type --- | --- | --- Flow | The flow | Flow Traces | The traces for this flow | Set of Trace TraceCount | The total number traces for this flow | int Retrieving the flow definition
result.Flow
docs/source/notebooks/forwarding.ipynb
batfish/pybatfish
apache-2.0
a5ce21ed4e4f59e3fb9ed05d195a64f4
Retrieving the detailed Trace information
len(result.Traces) result.Traces[0]
docs/source/notebooks/forwarding.ipynb
batfish/pybatfish
apache-2.0
f1042c1322839011dcdc37949ae1bea3
Evaluating the first Trace
result.Traces[0][0]
docs/source/notebooks/forwarding.ipynb
batfish/pybatfish
apache-2.0
b0a24f9750963cde08b232f0fc02acac
Retrieving the disposition of the first Trace
result.Traces[0][0].disposition
docs/source/notebooks/forwarding.ipynb
batfish/pybatfish
apache-2.0
92b75a93113a08c4327400613ceb41ca
Retrieving the first hop of the first Trace
result.Traces[0][0][0]
docs/source/notebooks/forwarding.ipynb
batfish/pybatfish
apache-2.0
856e0b952e1f1cafbb4424893dadf7d8
Retrieving the last hop of the first Trace
result.Traces[0][0][-1] bf.set_network('generate_questions') bf.set_snapshot('generate_questions')
docs/source/notebooks/forwarding.ipynb
batfish/pybatfish
apache-2.0
6f2efdb725c03b04aafc33853d515877
Bi-directional Traceroute Traces the path(s) for the specified flow, along with path(s) for reverse flows. This question performs a virtual traceroute in the network from a starting node. A destination IP and ingress (source) node must be specified. Other IP headers are given default values if unspecified. If the trace succeeds, a traceroute is performed in the reverse direction. Inputs Name | Description | Type | Optional | Default Value --- | --- | --- | --- | --- startLocation | Location (node and interface combination) to start tracing from. | LocationSpec | False | headers | Packet header constraints. | HeaderConstraints | False | maxTraces | Limit the number of traces returned. | int | True | ignoreFilters | If set, filters/ACLs encountered along the path are ignored. | bool | True | Invocation
result = bf.q.bidirectionalTraceroute(startLocation='@enter(as2border1[GigabitEthernet2/0])', headers=HeaderConstraints(dstIps='2.34.201.10', srcIps='8.8.8.8')).answer().frame()
docs/source/notebooks/forwarding.ipynb
batfish/pybatfish
apache-2.0
ce490c94257848d3073e6477d8104993