-
Notifications
You must be signed in to change notification settings - Fork 18
/
Copy pathplot_stock_market.py
310 lines (265 loc) · 10.1 KB
/
plot_stock_market.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
"""
see:
http://scikit-learn.org/stable/auto_examples/applications/plot_stock_market.html
http://scikit-learn.org/stable/_downloads/plot_stock_market.py
=======================================
Visualizing the stock market structure
=======================================
This example employs several unsupervised learning techniques to extract
the stock market structure from variations in historical quotes.
The quantity that we use is the daily variation in quote price: quotes
that are linked tend to cofluctuate during a day.
.. _stock_market:
Learning a graph structure
--------------------------
We use sparse inverse covariance estimation to find which quotes are
correlated conditionally on the others. Specifically, sparse inverse
covariance gives us a graph, that is a list of connection. For each
symbol, the symbols that it is connected too are those useful to explain
its fluctuations.
Clustering
----------
We use clustering to group together quotes that behave similarly. Here,
amongst the :ref:`various clustering techniques <clustering>` available
in the scikit-learn, we use :ref:`affinity_propagation` as it does
not enforce equal-size clusters, and it can choose automatically the
number of clusters from the data.
Note that this gives us a different indication than the graph, as the
graph reflects conditional relations between variables, while the
clustering reflects marginal properties: variables clustered together can
be considered as having a similar impact at the level of the full stock
market.
Embedding in 2D space
---------------------
For visualization purposes, we need to lay out the different symbols on a
2D canvas. For this we use :ref:`manifold` techniques to retrieve 2D
embedding.
Visualization
-------------
The output of the 3 models are combined in a 2D graph where nodes
represents the stocks and edges the:
- cluster labels are used to define the color of the nodes
- the sparse covariance model is used to display the strength of the edges
- the 2D embedding is used to position the nodes in the plan
This example has a fair amount of visualization-related code, as
visualization is crucial here to display the graph. One of the challenge
is to position the labels minimizing overlap. For this we use an
heuristic based on the direction of the nearest neighbor along each
axis.
"""
print(__doc__)
# Author: Gael Varoquaux [email protected]
# License: BSD 3 clause
import datetime
import numpy as np
import matplotlib.pyplot as plt
from matplotlib import finance
from matplotlib.collections import LineCollection
from sklearn import cluster, covariance, manifold
###############################################################################
# Retrieve the data from Internet
# Choose a time period reasonnably calm (not too long ago so that we get
# high-tech firms, and before the 2008 crash)
# d1 = datetime.datetime(2003, 1, 1)
# d2 = datetime.datetime(2008, 1, 1)
d1 = datetime.datetime(2009, 1, 1)
d2 = datetime.datetime(2015, 1, 1)
# 2015jan28: cls: anything after 2008 does NOT work!
# why ??? coz some of stocks are missing
# this error occurs when a ticker is not found within the date range (d1 to d2):
# Traceback (most recent call last):
# File "plot_stock_market.py", line 178, in <module>
# open = np.array([q.open for q in quotes]).astype(np.float)
# ValueError: setting an array element with a sequence.
# ... how to handle this, other than commenting out the ticker symbol ???
# kraft symbol has now changed from KFT to MDLZ in yahoo
# symbol_dict = {
# 'TOT': 'Total',
# 'XOM': 'Exxon',
# 'CVX': 'Chevron',
# 'COP': 'ConocoPhillips',
# 'VLO': 'Valero Energy',
# 'MSFT': 'Microsoft',
# 'IBM': 'IBM',
# 'TWX': 'Time Warner',
# 'CMCSA': 'Comcast',
# 'CVC': 'Cablevision',
# 'YHOO': 'Yahoo',
# 'DELL': 'Dell',
# 'HPQ': 'HP',
# 'AMZN': 'Amazon',
# 'TM': 'Toyota',
# 'CAJ': 'Canon',
# 'MTU': 'Mitsubishi',
# 'SNE': 'Sony',
# 'F': 'Ford',
# 'HMC': 'Honda',
# 'NAV': 'Navistar',
# 'NOC': 'Northrop Grumman',
# 'BA': 'Boeing',
# 'KO': 'Coca Cola',
# 'MMM': '3M',
# 'MCD': 'Mc Donalds',
# 'PEP': 'Pepsi',
# 'MDLZ': 'Kraft Foods',
# 'K': 'Kellogg',
# 'UN': 'Unilever',
# 'MAR': 'Marriott',
# 'PG': 'Procter Gamble',
# 'CL': 'Colgate-Palmolive',
# 'GE': 'General Electrics',
# 'WFC': 'Wells Fargo',
# 'JPM': 'JPMorgan Chase',
# 'AIG': 'AIG',
# 'AXP': 'American express',
# 'BAC': 'Bank of America',
# 'GS': 'Goldman Sachs',
# 'AAPL': 'Apple',
# 'SAP': 'SAP',
# 'CSCO': 'Cisco',
# 'TXN': 'Texas instruments',
# 'XRX': 'Xerox',
# 'LMT': 'Lookheed Martin',
# 'WMT': 'Wal-Mart',
# 'WAG': 'Walgreen',
# 'HD': 'Home Depot',
# 'GSK': 'GlaxoSmithKline',
# 'PFE': 'Pfizer',
# 'SNY': 'Sanofi-Aventis',
# 'NVS': 'Novartis',
# 'KMB': 'Kimberly-Clark',
# 'R': 'Ryder',
# 'GD': 'General Dynamics',
# 'RTN': 'Raytheon',
# 'CVS': 'CVS',
# 'CAT': 'Caterpillar',
# 'DD': 'DuPont de Nemours'}
symbol_dict = {
'TOT': 'Total',
'AAPL': 'Apple',
'MSFT': 'Microsoft',
'IBM': 'IBM',
'YHOO': 'Yahoo',
'AMZN': 'Amazon',
'KO': 'Coca Cola',
'PEP': 'Pepsi',
'MCD': 'Mc Donalds',
'YUM': 'Taco Bell',
'CMG': 'Chipotle Mexican Grill',
'WMT': 'Wal-Mart',
'HD': 'Home Depot',
'CVS': 'CVS'
}
symbols, names = np.array(list(symbol_dict.items())).T
print('_'*140 + "\n>>> quotes:")
for symbol in symbols:
try:
quote = finance.quotes_historical_yahoo_ochl(symbol, d1, d2, asobject=True)
print("\n%s %s:\n%s" % (symbol, symbol_dict[symbol], quote))
except Exception as e:
# this is usually a 404 error, coz of the date range (ie a stock may not have existed),
# so we could enter a zeroed entry ???
print('_'*140 + "\n")
cls1 = datetime.datetime(2009, 1, 1)
cls2 = datetime.datetime(2015, 1, 1)
symbol = 'YUM' # Taco Bell
# symbol = 'CMG' # Chipotle Mexican Grill
quote = finance.quotes_historical_yahoo_ochl(symbol, cls1, cls2, asobject=True)
print("\n%s %s\ntype(quote)=%s=\nquote:\n%s" % (symbol, symbol_dict[symbol], type(quote), quote))
# quotes = [finance.quotes_historical_yahoo(symbol, d1, d2, asobject=True) for symbol in symbols]
# quotes_historical_yahoo_ochl, which maintains the original argument order:
quotes = [finance.quotes_historical_yahoo_ochl(symbol, d1, d2, asobject=True) for symbol in symbols]
# quotes_historical_yahoo_ohlc, which uses the open-high-low-close order:
# quotes = [finance.quotes_historical_yahoo_ohlc(symbol, d1, d2, asobject=True) for symbol in symbols]
open = np.array([q.open for q in quotes]).astype(np.float)
close = np.array([q.close for q in quotes]).astype(np.float)
# The daily variations of the quotes are what carry most information
variation = close - open
# print("\n>>> variation:")
# print(variation)
# print('_'*140 + "\n")
###############################################################################
# Learn a graphical structure from the correlations
edge_model = covariance.GraphLassoCV()
# standardize the time series: using correlations rather than covariance
# is more efficient for structure recovery
X = variation.copy().T
X /= X.std(axis=0)
edge_model.fit(X)
###############################################################################
# Cluster using affinity propagation
_, labels = cluster.affinity_propagation(edge_model.covariance_)
n_labels = labels.max()
for i in range(n_labels + 1):
print('Cluster %i: %s' % ((i + 1), ', '.join(names[labels == i])))
###############################################################################
# Find a low-dimension embedding for visualization: find the best position of
# the nodes (the stocks) on a 2D plane
# We use a dense eigen_solver to achieve reproducibility (arpack is
# initiated with random vectors that we don't control). In addition, we
# use a large number of neighbors to capture the large-scale structure.
node_position_model = manifold.LocallyLinearEmbedding(
n_components=2, eigen_solver='dense', n_neighbors=6
)
embedding = node_position_model.fit_transform(X.T).T
###############################################################################
# Visualization
plt.figure(1, facecolor='w', figsize=(10, 8))
plt.clf()
ax = plt.axes([0., 0., 1., 1.])
plt.axis('off')
# Display a graph of the partial correlations
partial_correlations = edge_model.precision_.copy()
d = 1 / np.sqrt(np.diag(partial_correlations))
partial_correlations *= d
partial_correlations *= d[:, np.newaxis]
non_zero = (np.abs(np.triu(partial_correlations, k=1)) > 0.02)
# Plot the nodes using the coordinates of our embedding
plt.scatter(embedding[0], embedding[1], s=100 * d ** 2, c=labels, cmap=plt.cm.spectral)
# Plot the edges
start_idx, end_idx = np.where(non_zero)
#a sequence of (*line0*, *line1*, *line2*), where::
# linen = (x0, y0), (x1, y1), ... (xm, ym)
segments = [[embedding[:, start], embedding[:, stop]] for start, stop in zip(start_idx, end_idx)]
values = np.abs(partial_correlations[non_zero])
lc = LineCollection(
segments,
zorder=0, cmap=plt.cm.hot_r,
norm=plt.Normalize(0, .7 * values.max())
)
lc.set_array(values)
lc.set_linewidths(15 * values)
ax.add_collection(lc)
# Add a label to each node. The challenge here is that we want to
# position the labels to avoid overlap with other labels
for index, (name, label, (x, y)) in enumerate(zip(names, labels, embedding.T)):
dx = x - embedding[0]
dx[index] = 1
dy = y - embedding[1]
dy[index] = 1
this_dx = dx[np.argmin(np.abs(dy))]
this_dy = dy[np.argmin(np.abs(dx))]
if this_dx > 0:
horizontalalignment = 'left'
x = x + .002
else:
horizontalalignment = 'right'
x = x - .002
if this_dy > 0:
verticalalignment = 'bottom'
y = y + .002
else:
verticalalignment = 'top'
y = y - .002
plt.text(x, y, name, size=10,
horizontalalignment=horizontalalignment,
verticalalignment=verticalalignment,
bbox=dict(
facecolor='w',
edgecolor=plt.cm.spectral(label / float(n_labels)),
alpha=.6
)
)
plt.xlim(embedding[0].min() - .15 * embedding[0].ptp(), embedding[0].max() + .10 * embedding[0].ptp(),)
plt.ylim(embedding[1].min() - .03 * embedding[1].ptp(), embedding[1].max() + .03 * embedding[1].ptp())
plt.show()