Datasets:
Tasks:
Text Generation
Modalities:
Text
Formats:
parquet
Sub-tasks:
language-modeling
Languages:
code
Size:
100K - 1M
License:
markdown
stringlengths 0
37k
| code
stringlengths 1
33.3k
| path
stringlengths 8
215
| repo_name
stringlengths 6
77
| license
stringclasses 15
values |
---|---|---|---|---|
First, we'll download the dataset to our local machine. The data consists of characters rendered in a variety of fonts on a 28x28 image. The labels are limited to 'A' through 'J' (10 classes). The training set has about 500k and the testset 19000 labeled examples. Given these sizes, it should be possible to train models quickly on any machine. | url = 'https://commondatastorage.googleapis.com/books1000/'
last_percent_reported = None
data_root = '.' # Change me to store data elsewhere
def download_progress_hook(count, blockSize, totalSize):
"""A hook to report the progress of a download. This is mostly intended for users with
slow internet connections. Reports every 5% change in download progress.
"""
global last_percent_reported
percent = int(count * blockSize * 100 / totalSize)
if last_percent_reported != percent:
if percent % 5 == 0:
sys.stdout.write("%s%%" % percent)
sys.stdout.flush()
else:
sys.stdout.write(".")
sys.stdout.flush()
last_percent_reported = percent
def maybe_download(filename, expected_bytes, force=False):
"""Download a file if not present, and make sure it's the right size."""
dest_filename = os.path.join(data_root, filename)
if force or not os.path.exists(dest_filename):
print('Attempting to download:', filename)
filename, _ = urlretrieve(url + filename, dest_filename, reporthook=download_progress_hook)
print('\nDownload Complete!')
statinfo = os.stat(dest_filename)
if statinfo.st_size == expected_bytes:
print('Found and verified', dest_filename)
else:
raise Exception(
'Failed to verify ' + dest_filename + '. Can you get to it with a browser?')
return dest_filename
train_filename = maybe_download('notMNIST_large.tar.gz', 247336696)
test_filename = maybe_download('notMNIST_small.tar.gz', 8458043) | machine-learning/deep-learning/udacity/ud730/1_notmnist.ipynb | pk-ai/training | mit |
Extract the dataset from the compressed .tar.gz file.
This should give you a set of directories, labeled A through J. | num_classes = 10
np.random.seed(133)
def maybe_extract(filename, force=False):
root = os.path.splitext(os.path.splitext(filename)[0])[0] # remove .tar.gz
if os.path.isdir(root) and not force:
# You may override by setting force=True.
print('%s already present - Skipping extraction of %s.' % (root, filename))
else:
print('Extracting data for %s. This may take a while. Please wait.' % root)
tar = tarfile.open(filename)
sys.stdout.flush()
tar.extractall(data_root)
tar.close()
data_folders = [
os.path.join(root, d) for d in sorted(os.listdir(root))
if os.path.isdir(os.path.join(root, d))]
if len(data_folders) != num_classes:
raise Exception(
'Expected %d folders, one per class. Found %d instead.' % (
num_classes, len(data_folders)))
print(data_folders)
return data_folders
train_folders = maybe_extract(train_filename)
test_folders = maybe_extract(test_filename) | machine-learning/deep-learning/udacity/ud730/1_notmnist.ipynb | pk-ai/training | mit |
Problem 1
Let's take a peek at some of the data to make sure it looks sensible. Each exemplar should be an image of a character A through J rendered in a different font. Display a sample of the images that we just downloaded. Hint: you can use the package IPython.display. | # Solution for Problem 1
import random
print('Displaying images of train folders')
# Looping through train folders and displaying a random image of each folder
for path in train_folders:
image_file = os.path.join(path, random.choice(os.listdir(path)))
display(Image(filename=image_file))
print('Displaying images of test folders')
# Looping through train folders and displaying a random image of each folder
for path in test_folders:
image_file = os.path.join(path, random.choice(os.listdir(path)))
display(Image(filename=image_file)) | machine-learning/deep-learning/udacity/ud730/1_notmnist.ipynb | pk-ai/training | mit |
Now let's load the data in a more manageable format. Since, depending on your computer setup you might not be able to fit it all in memory, we'll load each class into a separate dataset, store them on disk and curate them independently. Later we'll merge them into a single dataset of manageable size.
We'll convert the entire dataset into a 3D array (image index, x, y) of floating point values, normalized to have approximately zero mean and standard deviation ~0.5 to make training easier down the road.
A few images might not be readable, we'll just skip them. | image_size = 28 # Pixel width and height.
pixel_depth = 255.0 # Number of levels per pixel.
def load_letter(folder, min_num_images):
"""Load the data for a single letter label."""
image_files = os.listdir(folder)
dataset = np.ndarray(shape=(len(image_files), image_size, image_size),
dtype=np.float32)
print(folder)
num_images = 0
for image in image_files:
image_file = os.path.join(folder, image)
try:
image_data = (ndimage.imread(image_file).astype(float) -
pixel_depth / 2) / pixel_depth
if image_data.shape != (image_size, image_size):
raise Exception('Unexpected image shape: %s' % str(image_data.shape))
dataset[num_images, :, :] = image_data
num_images = num_images + 1
except IOError as e:
print('Could not read:', image_file, ':', e, '- it\'s ok, skipping.')
dataset = dataset[0:num_images, :, :]
if num_images < min_num_images:
raise Exception('Many fewer images than expected: %d < %d' %
(num_images, min_num_images))
print('Full dataset tensor:', dataset.shape)
print('Mean:', np.mean(dataset))
print('Standard deviation:', np.std(dataset))
return dataset
def maybe_pickle(data_folders, min_num_images_per_class, force=False):
dataset_names = []
for folder in data_folders:
set_filename = folder + '.pickle'
dataset_names.append(set_filename)
if os.path.exists(set_filename) and not force:
# You may override by setting force=True.
print('%s already present - Skipping pickling.' % set_filename)
else:
print('Pickling %s.' % set_filename)
dataset = load_letter(folder, min_num_images_per_class)
try:
with open(set_filename, 'wb') as f:
pickle.dump(dataset, f, pickle.HIGHEST_PROTOCOL)
except Exception as e:
print('Unable to save data to', set_filename, ':', e)
return dataset_names
train_datasets = maybe_pickle(train_folders, 45000)
test_datasets = maybe_pickle(test_folders, 1800) | machine-learning/deep-learning/udacity/ud730/1_notmnist.ipynb | pk-ai/training | mit |
Problem 2
Let's verify that the data still looks good. Displaying a sample of the labels and images from the ndarray. Hint: you can use matplotlib.pyplot. | # Solution for Problem 2
def show_first_image(datasets):
for pickl in datasets:
print('Showing a first image from pickle ', pickl)
try:
with open(pickl, 'rb') as f:
letter_set = pickle.load(f)
plt.imshow(letter_set[0])
except Exception as e:
print('Unable to show image from pickle ', pickl, ':', e)
raise
print('From Training dataset')
show_first_image(train_datasets)
print('From Test Dataset')
show_first_image(test_datasets) | machine-learning/deep-learning/udacity/ud730/1_notmnist.ipynb | pk-ai/training | mit |
Problem 3
Another check: we expect the data to be balanced across classes. Verify that. | def show_dataset_shape(datasets):
for pickl in datasets:
try:
with open(pickl, 'rb') as f:
letter_set = pickle.load(f)
print('Shape of pickle ', pickl, 'is', np.shape(letter_set))
except Exception as e:
print('Unable to show image from pickle ', pickl, ':', e)
raise
print('Shape for Training set')
show_dataset_shape(train_datasets)
print('Shape for Test set')
show_dataset_shape(test_datasets) | machine-learning/deep-learning/udacity/ud730/1_notmnist.ipynb | pk-ai/training | mit |
Merge and prune the training data as needed. Depending on your computer setup, you might not be able to fit it all in memory, and you can tune train_size as needed. The labels will be stored into a separate array of integers 0 through 9.
Also create a validation dataset for hyperparameter tuning. | def make_arrays(nb_rows, img_size):
if nb_rows:
dataset = np.ndarray((nb_rows, img_size, img_size), dtype=np.float32)
labels = np.ndarray(nb_rows, dtype=np.int32)
else:
dataset, labels = None, None
return dataset, labels
def merge_datasets(pickle_files, train_size, valid_size=0):
num_classes = len(pickle_files)
valid_dataset, valid_labels = make_arrays(valid_size, image_size)
train_dataset, train_labels = make_arrays(train_size, image_size)
vsize_per_class = valid_size // num_classes
tsize_per_class = train_size // num_classes
start_v, start_t = 0, 0
end_v, end_t = vsize_per_class, tsize_per_class
end_l = vsize_per_class+tsize_per_class
for label, pickle_file in enumerate(pickle_files):
try:
with open(pickle_file, 'rb') as f:
letter_set = pickle.load(f)
# let's shuffle the letters to have random validation and training set
np.random.shuffle(letter_set)
if valid_dataset is not None:
valid_letter = letter_set[:vsize_per_class, :, :]
valid_dataset[start_v:end_v, :, :] = valid_letter
valid_labels[start_v:end_v] = label
start_v += vsize_per_class
end_v += vsize_per_class
train_letter = letter_set[vsize_per_class:end_l, :, :]
train_dataset[start_t:end_t, :, :] = train_letter
train_labels[start_t:end_t] = label
start_t += tsize_per_class
end_t += tsize_per_class
except Exception as e:
print('Unable to process data from', pickle_file, ':', e)
raise
return valid_dataset, valid_labels, train_dataset, train_labels
"""
train_size = 200000
valid_size = 10000
test_size = 10000
"""
train_size = 20000
valid_size = 1000
test_size = 1000
valid_dataset, valid_labels, train_dataset, train_labels = merge_datasets(
train_datasets, train_size, valid_size)
_, _, test_dataset, test_labels = merge_datasets(test_datasets, test_size)
print('Training:', train_dataset.shape, train_labels.shape)
print('Validation:', valid_dataset.shape, valid_labels.shape)
print('Testing:', test_dataset.shape, test_labels.shape) | machine-learning/deep-learning/udacity/ud730/1_notmnist.ipynb | pk-ai/training | mit |
Next, we'll randomize the data. It's important to have the labels well shuffled for the training and test distributions to match. | def randomize(dataset, labels):
permutation = np.random.permutation(labels.shape[0])
shuffled_dataset = dataset[permutation,:,:]
shuffled_labels = labels[permutation]
return shuffled_dataset, shuffled_labels
train_dataset, train_labels = randomize(train_dataset, train_labels)
test_dataset, test_labels = randomize(test_dataset, test_labels)
valid_dataset, valid_labels = randomize(valid_dataset, valid_labels) | machine-learning/deep-learning/udacity/ud730/1_notmnist.ipynb | pk-ai/training | mit |
Problem 4
Convince yourself that the data is still good after shuffling! | print('Printing Train, validation and test labels after shuffling')
def print_first_10_labels(labels):
printing_labels = []
for i in range(10):
printing_labels.append(labels[[i]])
print(printing_labels)
print_first_10_labels(train_labels)
print_first_10_labels(test_labels)
print_first_10_labels(valid_labels) | machine-learning/deep-learning/udacity/ud730/1_notmnist.ipynb | pk-ai/training | mit |
Finally, let's save the data for later reuse: | pickle_file = os.path.join(data_root, 'notMNIST.pickle')
try:
f = open(pickle_file, 'wb')
save = {
'train_dataset': train_dataset,
'train_labels': train_labels,
'valid_dataset': valid_dataset,
'valid_labels': valid_labels,
'test_dataset': test_dataset,
'test_labels': test_labels,
}
pickle.dump(save, f, pickle.HIGHEST_PROTOCOL)
f.close()
except Exception as e:
print('Unable to save data to', pickle_file, ':', e)
raise
statinfo = os.stat(pickle_file)
print('Compressed pickle size:', statinfo.st_size) | machine-learning/deep-learning/udacity/ud730/1_notmnist.ipynb | pk-ai/training | mit |
Problem 5
By construction, this dataset might contain a lot of overlapping samples, including training data that's also contained in the validation and test set! Overlap between training and test can skew the results if you expect to use your model in an environment where there is never an overlap, but are actually ok if you expect to see training samples recur when you use it.
Measure how much overlap there is between training, validation and test samples.
Optional questions:
- What about near duplicates between datasets? (images that are almost identical)
- Create a sanitized validation and test set, and compare your accuracy on those in subsequent assignments.
Problem 6
Let's get an idea of what an off-the-shelf classifier can give you on this data. It's always good to check that there is something to learn, and that it's a problem that is not so trivial that a canned solution solves it.
Train a simple model on this data using 50, 100, 1000 and 5000 training samples. Hint: you can use the LogisticRegression model from sklearn.linear_model.
Optional question: train an off-the-shelf model on all the data! | logreg_model_clf = LogisticRegression()
nsamples, nx, ny = train_dataset.shape
d2_train_dataset = train_dataset.reshape((nsamples,nx*ny))
logreg_model_clf.fit(d2_train_dataset, train_labels)
from sklearn.metrics import accuracy_score
nsamples, nx, ny = valid_dataset.shape
d2_valid_dataset = valid_dataset.reshape((nsamples,nx*ny))
print("validation accuracy,", accuracy_score(valid_labels, logreg_model_clf.predict(d2_valid_dataset)))
nsamples, nx, ny = test_dataset.shape
d2_train_dataset = test_dataset.reshape((nsamples,nx*ny))
print("test accuracy,", accuracy_score(test_labels, logreg_model_clf.predict(d2_train_dataset))) | machine-learning/deep-learning/udacity/ud730/1_notmnist.ipynb | pk-ai/training | mit |
Now the Hotels | url = 'http://www.bringfido.com/lodging/city/new_haven_ct_us'
r = Render(url)
result = r.frame.toHtml()
#QString should be converted to string before processed by lxml
formatted_result = str(result.toAscii())
tree = html.fromstring(formatted_result)
#Now using correct Xpath we are fetching URL of archives
archive_links = tree.xpath('//*[@id="results_list"]/div')
print(archive_links)
print('')
for lnk in archive_links:
print(lnk.xpath('div[2]/h1/a/text()')[0])
print(lnk.text_content())
print('*'*25)
| code/.ipynb_checkpoints/bf_qt_scraping-checkpoint.ipynb | mattgiguere/doglodge | mit |
Now Get the Links | links = []
for lnk in archive_links:
print(lnk.xpath('div/h1/a/@href')[0])
links.append(lnk.xpath('div/h1/a/@href')[0])
print('*'*25)
lnk.xpath('//*/div/h1/a/@href')[0]
links | code/.ipynb_checkpoints/bf_qt_scraping-checkpoint.ipynb | mattgiguere/doglodge | mit |
Loading Reviews
Next, we want to step through each page, and scrape the reviews for each hotel. | url_base = 'http://www.bringfido.com'
r.update_url(url_base+links[0])
result = r.frame.toHtml()
#QString should be converted to string before processed by lxml
formatted_result = str(result.toAscii())
tree = html.fromstring(formatted_result)
hotel_description = tree.xpath('//*[@class="body"]/text()')
details = tree.xpath('//*[@class="address"]/text()')
address = details[0]
csczip = details[1]
phone = details[2]
#Now using correct Xpath we are fetching URL of archives
reviews = tree.xpath('//*[@class="review_container"]')
texts = []
titles = []
authors = []
ratings = []
print(reviews)
print('')
for rev in reviews:
titles.append(rev.xpath('div/div[1]/text()')[0])
authors.append(rev.xpath('div/div[2]/text()')[0])
texts.append(rev.xpath('div/div[3]/text()')[0])
ratings.append(rev.xpath('div[2]/img/@src')[0].split('/')[-1][0:1])
print(rev.xpath('div[2]/img/@src')[0].split('/')[-1][0:1])
titles
authors
texts
ratings | code/.ipynb_checkpoints/bf_qt_scraping-checkpoint.ipynb | mattgiguere/doglodge | mit |
Load software and filenames definitions | from fretbursts import *
init_notebook()
from IPython.display import display | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Data folder: | data_dir = './data/singlespot/'
import os
data_dir = os.path.abspath(data_dir) + '/'
assert os.path.exists(data_dir), "Path '%s' does not exist." % data_dir | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
List of data files: | from glob import glob
file_list = sorted(f for f in glob(data_dir + '*.hdf5') if '_BKG' not in f)
## Selection for POLIMI 2012-11-26 datatset
labels = ['17d', '27d', '7d', '12d', '22d']
files_dict = {lab: fname for lab, fname in zip(labels, file_list)}
files_dict
data_id | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Data load
Initial loading of the data: | d = loader.photon_hdf5(filename=files_dict[data_id]) | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Load the leakage coefficient from disk: | leakage_coeff_fname = 'results/usALEX - leakage coefficient DexDem.csv'
leakage = np.loadtxt(leakage_coeff_fname)
print('Leakage coefficient:', leakage) | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Load the direct excitation coefficient ($d_{exAA}$) from disk: | dir_ex_coeff_fname = 'results/usALEX - direct excitation coefficient dir_ex_aa.csv'
dir_ex_aa = np.loadtxt(dir_ex_coeff_fname)
print('Direct excitation coefficient (dir_ex_aa):', dir_ex_aa) | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Load the gamma-factor ($\gamma$) from disk: | gamma_fname = 'results/usALEX - gamma factor - all-ph.csv'
gamma = np.loadtxt(gamma_fname)
print('Gamma-factor:', gamma) | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Update d with the correction coefficients: | d.leakage = leakage
d.dir_ex = dir_ex_aa
d.gamma = gamma | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Laser alternation selection
At this point we have only the timestamps and the detector numbers: | d.ph_times_t[0][:3], d.ph_times_t[0][-3:]#, d.det_t
print('First and last timestamps: {:10,} {:10,}'.format(d.ph_times_t[0][0], d.ph_times_t[0][-1]))
print('Total number of timestamps: {:10,}'.format(d.ph_times_t[0].size)) | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
We need to define some parameters: donor and acceptor ch, excitation period and donor and acceptor excitiations: | d.add(det_donor_accept=(0, 1), alex_period=4000, D_ON=(2850, 580), A_ON=(900, 2580), offset=0) | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
We should check if everithing is OK with an alternation histogram: | plot_alternation_hist(d) | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
If the plot looks good we can apply the parameters with: | loader.alex_apply_period(d)
print('D+A photons in D-excitation period: {:10,}'.format(d.D_ex[0].sum()))
print('D+A photons in A-excitation period: {:10,}'.format(d.A_ex[0].sum())) | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Measurements infos
All the measurement data is in the d variable. We can print it: | d | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Or check the measurements duration: | d.time_max | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Compute background
Compute the background using automatic threshold: | d.calc_bg(bg.exp_fit, time_s=60, tail_min_us='auto', F_bg=1.7)
dplot(d, timetrace_bg)
d.rate_m, d.rate_dd, d.rate_ad, d.rate_aa | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Burst search and selection | d.burst_search(L=10, m=10, F=7, ph_sel=Ph_sel('all'))
print(d.ph_sel)
dplot(d, hist_fret);
# if data_id in ['7d', '27d']:
# ds = d.select_bursts(select_bursts.size, th1=20)
# else:
# ds = d.select_bursts(select_bursts.size, th1=30)
ds = d.select_bursts(select_bursts.size, add_naa=False, th1=30)
n_bursts_all = ds.num_bursts[0]
def select_and_plot_ES(fret_sel, do_sel):
ds_fret= ds.select_bursts(select_bursts.ES, **fret_sel)
ds_do = ds.select_bursts(select_bursts.ES, **do_sel)
bpl.plot_ES_selection(ax, **fret_sel)
bpl.plot_ES_selection(ax, **do_sel)
return ds_fret, ds_do
ax = dplot(ds, hist2d_alex, S_max_norm=2, scatter_alpha=0.1)
if data_id == '7d':
fret_sel = dict(E1=0.60, E2=1.2, S1=0.2, S2=0.9, rect=False)
do_sel = dict(E1=-0.2, E2=0.5, S1=0.8, S2=2, rect=True)
ds_fret, ds_do = select_and_plot_ES(fret_sel, do_sel)
elif data_id == '12d':
fret_sel = dict(E1=0.30,E2=1.2,S1=0.131,S2=0.9, rect=False)
do_sel = dict(E1=-0.4, E2=0.4, S1=0.8, S2=2, rect=False)
ds_fret, ds_do = select_and_plot_ES(fret_sel, do_sel)
elif data_id == '17d':
fret_sel = dict(E1=0.01, E2=0.98, S1=0.14, S2=0.88, rect=False)
do_sel = dict(E1=-0.4, E2=0.4, S1=0.80, S2=2, rect=False)
ds_fret, ds_do = select_and_plot_ES(fret_sel, do_sel)
elif data_id == '22d':
fret_sel = dict(E1=-0.16, E2=0.6, S1=0.2, S2=0.80, rect=False)
do_sel = dict(E1=-0.2, E2=0.4, S1=0.85, S2=2, rect=True)
ds_fret, ds_do = select_and_plot_ES(fret_sel, do_sel)
elif data_id == '27d':
fret_sel = dict(E1=-0.1, E2=0.5, S1=0.2, S2=0.82, rect=False)
do_sel = dict(E1=-0.2, E2=0.4, S1=0.88, S2=2, rect=True)
ds_fret, ds_do = select_and_plot_ES(fret_sel, do_sel)
n_bursts_do = ds_do.num_bursts[0]
n_bursts_fret = ds_fret.num_bursts[0]
n_bursts_do, n_bursts_fret
d_only_frac = 1.*n_bursts_do/(n_bursts_do + n_bursts_fret)
print('D-only fraction:', d_only_frac)
dplot(ds_fret, hist2d_alex, scatter_alpha=0.1);
dplot(ds_do, hist2d_alex, S_max_norm=2, scatter=False); | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Donor Leakage fit | bandwidth = 0.03
E_range_do = (-0.1, 0.15)
E_ax = np.r_[-0.2:0.401:0.0002]
E_pr_do_kde = bext.fit_bursts_kde_peak(ds_do, bandwidth=bandwidth, weights='size',
x_range=E_range_do, x_ax=E_ax, save_fitter=True)
mfit.plot_mfit(ds_do.E_fitter, plot_kde=True, bins=np.r_[E_ax.min(): E_ax.max(): bandwidth])
plt.xlim(-0.3, 0.5)
print("%s: E_peak = %.2f%%" % (ds.ph_sel, E_pr_do_kde*100)) | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Burst sizes | nt_th1 = 50
dplot(ds_fret, hist_size, which='all', add_naa=False)
xlim(-0, 250)
plt.axvline(nt_th1)
Th_nt = np.arange(35, 120)
nt_th = np.zeros(Th_nt.size)
for i, th in enumerate(Th_nt):
ds_nt = ds_fret.select_bursts(select_bursts.size, th1=th)
nt_th[i] = (ds_nt.nd[0] + ds_nt.na[0]).mean() - th
plt.figure()
plot(Th_nt, nt_th)
plt.axvline(nt_th1)
nt_mean = nt_th[np.where(Th_nt == nt_th1)][0]
nt_mean | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Fret fit
Max position of the Kernel Density Estimation (KDE): | E_pr_fret_kde = bext.fit_bursts_kde_peak(ds_fret, bandwidth=bandwidth, weights='size')
E_fitter = ds_fret.E_fitter
E_fitter.histogram(bins=np.r_[-0.1:1.1:0.03])
E_fitter.fit_histogram(mfit.factory_gaussian(center=0.5))
E_fitter.fit_res[0].params.pretty_print()
fig, ax = plt.subplots(1, 2, figsize=(14, 4.5))
mfit.plot_mfit(E_fitter, ax=ax[0])
mfit.plot_mfit(E_fitter, plot_model=False, plot_kde=True, ax=ax[1])
print('%s\nKDE peak %.2f ' % (ds_fret.ph_sel, E_pr_fret_kde*100))
display(E_fitter.params*100) | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Weighted mean of $E$ of each burst: | ds_fret.fit_E_m(weights='size') | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Gaussian fit (no weights): | ds_fret.fit_E_generic(fit_fun=bl.gaussian_fit_hist, bins=np.r_[-0.1:1.1:0.03], weights=None) | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Gaussian fit (using burst size as weights): | ds_fret.fit_E_generic(fit_fun=bl.gaussian_fit_hist, bins=np.r_[-0.1:1.1:0.005], weights='size')
E_kde_w = E_fitter.kde_max_pos[0]
E_gauss_w = E_fitter.params.loc[0, 'center']
E_gauss_w_sig = E_fitter.params.loc[0, 'sigma']
E_gauss_w_err = float(E_gauss_w_sig/np.sqrt(ds_fret.num_bursts[0]))
E_gauss_w_fiterr = E_fitter.fit_res[0].params['center'].stderr
E_kde_w, E_gauss_w, E_gauss_w_sig, E_gauss_w_err, E_gauss_w_fiterr | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Stoichiometry fit
Max position of the Kernel Density Estimation (KDE): | S_pr_fret_kde = bext.fit_bursts_kde_peak(ds_fret, burst_data='S', bandwidth=0.03) #weights='size', add_naa=True)
S_fitter = ds_fret.S_fitter
S_fitter.histogram(bins=np.r_[-0.1:1.1:0.03])
S_fitter.fit_histogram(mfit.factory_gaussian(), center=0.5)
fig, ax = plt.subplots(1, 2, figsize=(14, 4.5))
mfit.plot_mfit(S_fitter, ax=ax[0])
mfit.plot_mfit(S_fitter, plot_model=False, plot_kde=True, ax=ax[1])
print('%s\nKDE peak %.2f ' % (ds_fret.ph_sel, S_pr_fret_kde*100))
display(S_fitter.params*100)
S_kde = S_fitter.kde_max_pos[0]
S_gauss = S_fitter.params.loc[0, 'center']
S_gauss_sig = S_fitter.params.loc[0, 'sigma']
S_gauss_err = float(S_gauss_sig/np.sqrt(ds_fret.num_bursts[0]))
S_gauss_fiterr = S_fitter.fit_res[0].params['center'].stderr
S_kde, S_gauss, S_gauss_sig, S_gauss_err, S_gauss_fiterr | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
The Maximum likelihood fit for a Gaussian population is the mean: | S = ds_fret.S[0]
S_ml_fit = (S.mean(), S.std())
S_ml_fit | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Computing the weighted mean and weighted standard deviation we get: | weights = bl.fret_fit.get_weights(ds_fret.nd[0], ds_fret.na[0], weights='size', naa=ds_fret.naa[0], gamma=1.)
S_mean = np.dot(weights, S)/weights.sum()
S_std_dev = np.sqrt(
np.dot(weights, (S - S_mean)**2)/weights.sum())
S_wmean_fit = [S_mean, S_std_dev]
S_wmean_fit | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Save data to file | sample = data_id | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
The following string contains the list of variables to be saved. When saving, the order of the variables is preserved. | variables = ('sample n_bursts_all n_bursts_do n_bursts_fret '
'E_kde_w E_gauss_w E_gauss_w_sig E_gauss_w_err E_gauss_w_fiterr '
'S_kde S_gauss S_gauss_sig S_gauss_err S_gauss_fiterr '
'E_pr_do_kde nt_mean\n') | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
This is just a trick to format the different variables: | variables_csv = variables.replace(' ', ',')
fmt_float = '{%s:.6f}'
fmt_int = '{%s:d}'
fmt_str = '{%s}'
fmt_dict = {**{'sample': fmt_str},
**{k: fmt_int for k in variables.split() if k.startswith('n_bursts')}}
var_dict = {name: eval(name) for name in variables.split()}
var_fmt = ', '.join([fmt_dict.get(name, fmt_float) % name for name in variables.split()]) + '\n'
data_str = var_fmt.format(**var_dict)
print(variables_csv)
print(data_str)
# NOTE: The file name should be the notebook name but with .csv extension
with open('results/usALEX-5samples-E-corrected-all-ph.csv', 'a') as f:
f.seek(0, 2)
if f.tell() == 0:
f.write(variables_csv)
f.write(data_str) | out_notebooks/usALEX-5samples-E-corrected-all-ph-out-12d.ipynb | tritemio/multispot_paper | mit |
Data folder: | data_dir = './data/singlespot/' | out_notebooks/usALEX-5samples-PR-raw-dir_ex_aa-fit-out-AexAem-17d.ipynb | tritemio/multispot_paper | mit |
Check that the folder exists: | import os
data_dir = os.path.abspath(data_dir) + '/'
assert os.path.exists(data_dir), "Path '%s' does not exist." % data_dir | out_notebooks/usALEX-5samples-PR-raw-dir_ex_aa-fit-out-AexAem-17d.ipynb | tritemio/multispot_paper | mit |
List of data files in data_dir: | from glob import glob
file_list = sorted(f for f in glob(data_dir + '*.hdf5') if '_BKG' not in f)
file_list
## Selection for POLIMI 2012-12-6 dataset
# file_list.pop(2)
# file_list = file_list[1:-2]
# display(file_list)
# labels = ['22d', '27d', '17d', '12d', '7d']
## Selection for P.E. 2012-12-6 dataset
# file_list.pop(1)
# file_list = file_list[:-1]
# display(file_list)
# labels = ['22d', '27d', '17d', '12d', '7d']
## Selection for POLIMI 2012-11-26 datatset
labels = ['17d', '27d', '7d', '12d', '22d']
files_dict = {lab: fname for lab, fname in zip(labels, file_list)}
files_dict
ph_sel_map = {'all-ph': Ph_sel('all'), 'AexAem': Ph_sel(Aex='Aem')}
ph_sel = ph_sel_map[ph_sel_name]
data_id, ph_sel_name | out_notebooks/usALEX-5samples-PR-raw-dir_ex_aa-fit-out-AexAem-17d.ipynb | tritemio/multispot_paper | mit |
Laser alternation selection
At this point we have only the timestamps and the detector numbers: | d.ph_times_t, d.det_t | out_notebooks/usALEX-5samples-PR-raw-dir_ex_aa-fit-out-AexAem-17d.ipynb | tritemio/multispot_paper | mit |
If the plot looks good we can apply the parameters with: | loader.alex_apply_period(d) | out_notebooks/usALEX-5samples-PR-raw-dir_ex_aa-fit-out-AexAem-17d.ipynb | tritemio/multispot_paper | mit |
Burst search and selection | from mpl_toolkits.axes_grid1 import AxesGrid
import lmfit
print('lmfit version:', lmfit.__version__)
assert d.dir_ex == 0
assert d.leakage == 0
d.burst_search(m=10, F=6, ph_sel=ph_sel)
print(d.ph_sel, d.num_bursts)
ds_sa = d.select_bursts(select_bursts.naa, th1=30)
ds_sa.num_bursts | out_notebooks/usALEX-5samples-PR-raw-dir_ex_aa-fit-out-AexAem-17d.ipynb | tritemio/multispot_paper | mit |
Preliminary selection and plots | mask = (d.naa[0] - np.abs(d.na[0] + d.nd[0])) > 30
ds_saw = d.select_bursts_mask_apply([mask])
ds_sas0 = ds_sa.select_bursts(select_bursts.S, S2=0.10)
ds_sas = ds_sa.select_bursts(select_bursts.S, S2=0.15)
ds_sas2 = ds_sa.select_bursts(select_bursts.S, S2=0.20)
ds_sas3 = ds_sa.select_bursts(select_bursts.S, S2=0.25)
ds_st = d.select_bursts(select_bursts.size, add_naa=True, th1=30)
ds_sas.num_bursts
dx = ds_sas0
size = dx.na[0] + dx.nd[0]
s_hist, s_bins = np.histogram(size, bins=np.r_[-15 : 25 : 1], density=True)
s_ax = s_bins[:-1] + 0.5*(s_bins[1] - s_bins[0])
plot(s_ax, s_hist, '-o', alpha=0.5)
dx = ds_sas
size = dx.na[0] + dx.nd[0]
s_hist, s_bins = np.histogram(size, bins=np.r_[-15 : 25 : 1], density=True)
s_ax = s_bins[:-1] + 0.5*(s_bins[1] - s_bins[0])
plot(s_ax, s_hist, '-o', alpha=0.5)
dx = ds_sas2
size = dx.na[0] + dx.nd[0]
s_hist, s_bins = np.histogram(size, bins=np.r_[-15 : 25 : 1], density=True)
s_ax = s_bins[:-1] + 0.5*(s_bins[1] - s_bins[0])
plot(s_ax, s_hist, '-o', alpha=0.5)
dx = ds_sas3
size = dx.na[0] + dx.nd[0]
s_hist, s_bins = np.histogram(size, bins=np.r_[-15 : 25 : 1], density=True)
s_ax = s_bins[:-1] + 0.5*(s_bins[1] - s_bins[0])
plot(s_ax, s_hist, '-o', alpha=0.5)
plt.title('(nd + na) for A-only population using different S cutoff');
dx = ds_sa
alex_jointplot(dx);
dplot(ds_sa, hist_S) | out_notebooks/usALEX-5samples-PR-raw-dir_ex_aa-fit-out-AexAem-17d.ipynb | tritemio/multispot_paper | mit |
A-direct excitation fitting
To extract the A-direct excitation coefficient we need to fit the
S values for the A-only population.
The S value for the A-only population is fitted with different methods:
- Histogram git with 2 Gaussians or with 2 asymmetric Gaussians
(an asymmetric Gaussian has right- and left-side of the peak
decreasing according to different sigmas).
- KDE maximum
In the following we apply these methods using different selection
or weighting schemes to reduce amount of FRET population and make
fitting of the A-only population easier.
Even selection
Here A-only and FRET population are evenly selected. | dx = ds_sa
bin_width = 0.03
bandwidth = 0.03
bins = np.r_[-0.2 : 1 : bin_width]
x_kde = np.arange(bins.min(), bins.max(), 0.0002)
## Weights
weights = None
## Histogram fit
fitter_g = mfit.MultiFitter(dx.S)
fitter_g.histogram(bins=np.r_[-0.2 : 1.2 : bandwidth])
fitter_g.fit_histogram(model = mfit.factory_two_gaussians(p1_center=0.1, p2_center=0.4))
S_hist_orig = fitter_g.hist_pdf
S_2peaks = fitter_g.params.loc[0, 'p1_center']
dir_ex_S2p = S_2peaks/(1 - S_2peaks)
print('Fitted direct excitation (na/naa) [2-Gauss]:', dir_ex_S2p)
## KDE
fitter_g.calc_kde(bandwidth=bandwidth)
fitter_g.find_kde_max(x_kde, xmin=0, xmax=0.15)
S_peak = fitter_g.kde_max_pos[0]
dir_ex_S_kde = S_peak/(1 - S_peak)
print('Fitted direct excitation (na/naa) [KDE]: ', dir_ex_S_kde)
fig, ax = plt.subplots(1, 2, figsize=(14, 4.5))
mfit.plot_mfit(fitter_g, ax=ax[0])
ax[0].set_title('2-Gaussians fit (S_fit = %.2f %%)' % (S_2peaks*100))
mfit.plot_mfit(fitter_g, ax=ax[1], plot_model=False, plot_kde=True)
ax[1].set_title('KDE fit (S_fit = %.2f %%)' % (S_peak*100));
## 2-Asym-Gaussian
fitter_ag = mfit.MultiFitter(dx.S)
fitter_ag.histogram(bins=np.r_[-0.2 : 1.2 : bandwidth])
fitter_ag.fit_histogram(model = mfit.factory_two_asym_gaussians(p1_center=0.1, p2_center=0.4))
#print(fitter_ag.fit_obj[0].model.fit_report())
S_2peaks_a = fitter_ag.params.loc[0, 'p1_center']
dir_ex_S2pa = S_2peaks_a/(1 - S_2peaks_a)
print('Fitted direct excitation (na/naa) [2-Gauss]:', dir_ex_S2pa)
fig, ax = plt.subplots(1, 2, figsize=(14, 4.5))
mfit.plot_mfit(fitter_g, ax=ax[0])
ax[0].set_title('2-Gaussians fit (S_fit = %.2f %%)' % (S_2peaks*100))
mfit.plot_mfit(fitter_ag, ax=ax[1])
ax[1].set_title('2-Asym-Gaussians fit (S_fit = %.2f %%)' % (S_2peaks_a*100)); | out_notebooks/usALEX-5samples-PR-raw-dir_ex_aa-fit-out-AexAem-17d.ipynb | tritemio/multispot_paper | mit |
Zero threshold on nd
Select bursts with:
$$n_d < 0$$. | dx = ds_sa.select_bursts(select_bursts.nd, th1=-100, th2=0)
fitter = bext.bursts_fitter(dx, 'S')
fitter.fit_histogram(model = mfit.factory_gaussian(center=0.1))
S_1peaks_th = fitter.params.loc[0, 'center']
dir_ex_S1p = S_1peaks_th/(1 - S_1peaks_th)
print('Fitted direct excitation (na/naa) [2-Gauss]:', dir_ex_S1p)
mfit.plot_mfit(fitter)
plt.xlim(-0.1, 0.6) | out_notebooks/usALEX-5samples-PR-raw-dir_ex_aa-fit-out-AexAem-17d.ipynb | tritemio/multispot_paper | mit |
Selection 1
Bursts are weighted using $w = f(S)$, where the function $f(S)$ is a
Gaussian fitted to the $S$ histogram of the FRET population. | dx = ds_sa
## Weights
weights = 1 - mfit.gaussian(dx.S[0], fitter_g.params.loc[0, 'p2_center'], fitter_g.params.loc[0, 'p2_sigma'])
weights[dx.S[0] >= fitter_g.params.loc[0, 'p2_center']] = 0
## Histogram fit
fitter_w1 = mfit.MultiFitter(dx.S)
fitter_w1.weights = [weights]
fitter_w1.histogram(bins=np.r_[-0.2 : 1.2 : bandwidth])
fitter_w1.fit_histogram(model = mfit.factory_two_gaussians(p1_center=0.1, p2_center=0.4))
S_2peaks_w1 = fitter_w1.params.loc[0, 'p1_center']
dir_ex_S2p_w1 = S_2peaks_w1/(1 - S_2peaks_w1)
print('Fitted direct excitation (na/naa) [2-Gauss]:', dir_ex_S2p_w1)
## KDE
fitter_w1.calc_kde(bandwidth=bandwidth)
fitter_w1.find_kde_max(x_kde, xmin=0, xmax=0.15)
S_peak_w1 = fitter_w1.kde_max_pos[0]
dir_ex_S_kde_w1 = S_peak_w1/(1 - S_peak_w1)
print('Fitted direct excitation (na/naa) [KDE]: ', dir_ex_S_kde_w1)
def plot_weights(x, weights, ax):
ax2 = ax.twinx()
x_sort = x.argsort()
ax2.plot(x[x_sort], weights[x_sort], color='k', lw=4, alpha=0.4)
ax2.set_ylabel('Weights');
fig, ax = plt.subplots(1, 2, figsize=(14, 4.5))
mfit.plot_mfit(fitter_w1, ax=ax[0])
mfit.plot_mfit(fitter_g, ax=ax[0], plot_model=False, plot_kde=False)
plot_weights(dx.S[0], weights, ax=ax[0])
ax[0].set_title('2-Gaussians fit (S_fit = %.2f %%)' % (S_2peaks_w1*100))
mfit.plot_mfit(fitter_w1, ax=ax[1], plot_model=False, plot_kde=True)
mfit.plot_mfit(fitter_g, ax=ax[1], plot_model=False, plot_kde=False)
plot_weights(dx.S[0], weights, ax=ax[1])
ax[1].set_title('KDE fit (S_fit = %.2f %%)' % (S_peak_w1*100)); | out_notebooks/usALEX-5samples-PR-raw-dir_ex_aa-fit-out-AexAem-17d.ipynb | tritemio/multispot_paper | mit |
Selection 2
Bursts are here weighted using weights $w$:
$$w = n_{aa} - |n_a + n_d|$$ | ## Weights
sizes = dx.nd[0] + dx.na[0] #- dir_ex_S_kde_w3*dx.naa[0]
weights = dx.naa[0] - abs(sizes)
weights[weights < 0] = 0
## Histogram
fitter_w4 = mfit.MultiFitter(dx.S)
fitter_w4.weights = [weights]
fitter_w4.histogram(bins=np.r_[-0.2 : 1.2 : bandwidth])
fitter_w4.fit_histogram(model = mfit.factory_two_gaussians(p1_center=0.1, p2_center=0.4))
S_2peaks_w4 = fitter_w4.params.loc[0, 'p1_center']
dir_ex_S2p_w4 = S_2peaks_w4/(1 - S_2peaks_w4)
print('Fitted direct excitation (na/naa) [2-Gauss]:', dir_ex_S2p_w4)
## KDE
fitter_w4.calc_kde(bandwidth=bandwidth)
fitter_w4.find_kde_max(x_kde, xmin=0, xmax=0.15)
S_peak_w4 = fitter_w4.kde_max_pos[0]
dir_ex_S_kde_w4 = S_peak_w4/(1 - S_peak_w4)
print('Fitted direct excitation (na/naa) [KDE]: ', dir_ex_S_kde_w4)
fig, ax = plt.subplots(1, 2, figsize=(14, 4.5))
mfit.plot_mfit(fitter_w4, ax=ax[0])
mfit.plot_mfit(fitter_g, ax=ax[0], plot_model=False, plot_kde=False)
#plot_weights(dx.S[0], weights, ax=ax[0])
ax[0].set_title('2-Gaussians fit (S_fit = %.2f %%)' % (S_2peaks_w4*100))
mfit.plot_mfit(fitter_w4, ax=ax[1], plot_model=False, plot_kde=True)
mfit.plot_mfit(fitter_g, ax=ax[1], plot_model=False, plot_kde=False)
#plot_weights(dx.S[0], weights, ax=ax[1])
ax[1].set_title('KDE fit (S_fit = %.2f %%)' % (S_peak_w4*100)); | out_notebooks/usALEX-5samples-PR-raw-dir_ex_aa-fit-out-AexAem-17d.ipynb | tritemio/multispot_paper | mit |
Selection 3
Bursts are here selected according to:
$$n_{aa} - |n_a + n_d| > 30$$ | mask = (d.naa[0] - np.abs(d.na[0] + d.nd[0])) > 30
ds_saw = d.select_bursts_mask_apply([mask])
print(ds_saw.num_bursts)
dx = ds_saw
## Weights
weights = None
## 2-Gaussians
fitter_w5 = mfit.MultiFitter(dx.S)
fitter_w5.histogram(bins=np.r_[-0.2 : 1.2 : bandwidth])
fitter_w5.fit_histogram(model = mfit.factory_two_gaussians(p1_center=0.1, p2_center=0.4))
S_2peaks_w5 = fitter_w5.params.loc[0, 'p1_center']
dir_ex_S2p_w5 = S_2peaks_w5/(1 - S_2peaks_w5)
print('Fitted direct excitation (na/naa) [2-Gauss]:', dir_ex_S2p_w5)
## KDE
fitter_w5.calc_kde(bandwidth=bandwidth)
fitter_w5.find_kde_max(x_kde, xmin=0, xmax=0.15)
S_peak_w5 = fitter_w5.kde_max_pos[0]
S_2peaks_w5_fiterr = fitter_w5.fit_res[0].params['p1_center'].stderr
dir_ex_S_kde_w5 = S_peak_w5/(1 - S_peak_w5)
print('Fitted direct excitation (na/naa) [KDE]: ', dir_ex_S_kde_w5)
## 2-Asym-Gaussians
fitter_w5a = mfit.MultiFitter(dx.S)
fitter_w5a.histogram(bins=np.r_[-0.2 : 1.2 : bandwidth])
fitter_w5a.fit_histogram(model = mfit.factory_two_asym_gaussians(p1_center=0.05, p2_center=0.3))
S_2peaks_w5a = fitter_w5a.params.loc[0, 'p1_center']
dir_ex_S2p_w5a = S_2peaks_w5a/(1 - S_2peaks_w5a)
#print(fitter_w5a.fit_obj[0].model.fit_report(min_correl=0.5))
print('Fitted direct excitation (na/naa) [2-Asym-Gauss]:', dir_ex_S2p_w5a)
fig, ax = plt.subplots(1, 3, figsize=(19, 4.5))
mfit.plot_mfit(fitter_w5, ax=ax[0])
mfit.plot_mfit(fitter_g, ax=ax[0], plot_model=False, plot_kde=False)
ax[0].set_title('2-Gaussians fit (S_fit = %.2f %%)' % (S_2peaks_w5*100))
mfit.plot_mfit(fitter_w5, ax=ax[1], plot_model=False, plot_kde=True)
mfit.plot_mfit(fitter_g, ax=ax[1], plot_model=False, plot_kde=False)
ax[1].set_title('KDE fit (S_fit = %.2f %%)' % (S_peak_w5*100));
mfit.plot_mfit(fitter_w5a, ax=ax[2])
mfit.plot_mfit(fitter_g, ax=ax[2], plot_model=False, plot_kde=False)
ax[2].set_title('2-Asym-Gaussians fit (S_fit = %.2f %%)' % (S_2peaks_w5a*100)); | out_notebooks/usALEX-5samples-PR-raw-dir_ex_aa-fit-out-AexAem-17d.ipynb | tritemio/multispot_paper | mit |
Save data to file | sample = data_id
n_bursts_aa = ds_sas.num_bursts[0] | out_notebooks/usALEX-5samples-PR-raw-dir_ex_aa-fit-out-AexAem-17d.ipynb | tritemio/multispot_paper | mit |
The following string contains the list of variables to be saved. When saving, the order of the variables is preserved. | variables = ('sample n_bursts_aa dir_ex_S1p dir_ex_S_kde dir_ex_S2p dir_ex_S2pa '
'dir_ex_S2p_w1 dir_ex_S_kde_w1 dir_ex_S_kde_w4 dir_ex_S_kde_w5 dir_ex_S2p_w5 dir_ex_S2p_w5a '
'S_2peaks_w5 S_2peaks_w5_fiterr\n') | out_notebooks/usALEX-5samples-PR-raw-dir_ex_aa-fit-out-AexAem-17d.ipynb | tritemio/multispot_paper | mit |
This is just a trick to format the different variables: | variables_csv = variables.replace(' ', ',')
fmt_float = '{%s:.6f}'
fmt_int = '{%s:d}'
fmt_str = '{%s}'
fmt_dict = {**{'sample': fmt_str},
**{k: fmt_int for k in variables.split() if k.startswith('n_bursts')}}
var_dict = {name: eval(name) for name in variables.split()}
var_fmt = ', '.join([fmt_dict.get(name, fmt_float) % name for name in variables.split()]) + '\n'
data_str = var_fmt.format(**var_dict)
print(variables_csv)
print(data_str)
# NOTE: The file name should be the notebook name but with .csv extension
with open('results/usALEX-5samples-PR-raw-dir_ex_aa-fit-%s.csv' % ph_sel_name, 'a') as f:
f.seek(0, 2)
if f.tell() == 0:
f.write(variables_csv)
f.write(data_str) | out_notebooks/usALEX-5samples-PR-raw-dir_ex_aa-fit-out-AexAem-17d.ipynb | tritemio/multispot_paper | mit |
1. Get Arxiv data about machine learning
Write a APi querier and extract papers with the terms machine learning or artificial intelligence. Get 2000 results... and play nice! | class Arxiv_querier():
'''
This class takes as an input a query and the number of results, and returns all the parsed results.
Includes routines to deal with multiple pages of results.
'''
def __init__(self,base_url="http://export.arxiv.org/api/query?"):
'''
Initialise
'''
self.base_url = base_url
def query(self,query_string,max_results=100,wait_time=3):
'''
Query the base url
'''
#Attribute query string
#Load base URL
base_url = self.base_url
#Prepare query string
processed_query = re.sub(' ','+',query_string)
self.query_string="_".join(query_string.split(" "))
start=0
pages = 0
#Run the query and store results for as long as the number of results is bigger than the max results
keep_running = True
result_store = []
while keep_running==True:
pages +=1
print(pages)
#Query url (NB the start arg, which will change as we go through different
#pages)
query_url = base_url+'search_query=all:{q}&start={s}&max_results={max_res}'.format(
q=processed_query,s=start,max_res=max_results)
#Download
source = requests.get(query_url)
#Parse the xml and get the entries (papers)
parsed = feedparser.parse(source.content)
#Extract entries
entries = parsed['entries']
#If the number of entries is bigger than the maximum number of results
#this means we need to go to another page. We do that by offseting the
#start with max results
result_store.append(entries)
if len(entries)==max_results:
start+=max_results
#If we have less than max results this means we have run out of
#results and we toggle the keep_running switch off.
if len(entries)<max_results:
keep_running=False
time.sleep(wait_time)
#Save results in a flat list
self.entry_results = [x for el in result_store for x in el]
def extract_data(self):
'''
Here we extract data from the entries
'''
#Load entries
entries = self.entry_results
#Create df
output = pd.concat([pd.DataFrame({
'query':self.query_string,
'id':x['id'],
'link':x['link'],
'title':x['title'],
'authors':", ".join([el['name'] for el in x['authors']]),
'summary':x['summary'],
'updated':x['updated'],
'published':x['published'],
'category':x['arxiv_primary_category']['term'],
'pdf':str([el['href'] for el in x['links'] if el['type']=='application/pdf'][0]
)},index=[0]) for x in entries]).reset_index(drop=True)
output['year_published'] = [x.split("-")[0] for x in output['published']]
self.output_df = output
query_terms = ['artificial intelligence','machine learning','deep learning']
#There are some inconsistencies in the number of results so we run the query three times for each
#term and remove duplicated results
def extract_arxiv_data(term,max_results=1000,wait_time=10, tests=3):
'''
This function initialises the Arxiv_querier class, extracts the data and outputs it
'''
print(term)
collected = []
#We collect the data thrice
for i in np.arange(tests):
print('run'+ ' ' +str(i))
initialised = Arxiv_querier()
initialised.query(term,max_results,wait_time)
initialised.extract_data()
out = initialised.output_df
collected.append(out)
#We concatenate the dfs and remove the duplicates.
output = pd.concat(collected)
output_no_dupes = output.drop_duplicates('id')
#Return both
return([output,output_no_dupes])
arxiv_ai_results_three = [extract_arxiv_data(term=q) for q in query_terms]
all_papers = pd.concat([x[1] for x in arxiv_ai_results_three]).drop_duplicates('id').reset_index(drop=True)
print(all_papers.shape)
all_papers.head()
all_papers.to_csv(int_data+'/{today}_ai_papers.csv'.format(today=today_str),index=False) | notebooks/ml_topic_analysis_exploration.ipynb | Juan-Mateos/coll_int_ai_case | mit |
2. Some exploratory analysis | from nltk.corpus import stopwords
from nltk.tokenize import word_tokenize, sent_tokenize, RegexpTokenizer, PunktSentenceTokenizer
from nltk.stem import WordNetLemmatizer, SnowballStemmer, PorterStemmer
import scipy
import ast
import string as st
from bs4 import BeautifulSoup
import gensim
from gensim.models.coherencemodel import CoherenceModel
from sklearn.feature_extraction.text import TfidfVectorizer
from itertools import product
stopwords_c = stopwords.words('english')
stemmer = PorterStemmer()
lemmatizer= WordNetLemmatizer()
#Read papers
all_papers = pd.read_csv(int_data+'/19_8_2017_ai_papers.csv'.format(today=today_str))
#Let's begin by looking at years
#When where they published?
#Year distribution
year_pubs = all_papers['year_published'].value_counts()
year_pubs.index = [int(x) for x in year_pubs.index]
fig,ax = plt.subplots(figsize=(10,5))
year_pubs_sorted = year_pubs[sorted(year_pubs.index)]
year_pubs_subset = year_pubs_sorted[year_pubs_sorted.index>1991]
ax.plot(np.arange(1993,2018),year_pubs_subset.cumsum(),color='red')
ax.bar(np.arange(1993,2018),year_pubs_subset)
ax.hlines(xmin=1993,xmax=2017,y=[10000,20000,30000,40000],colors='green',linestyles='dashed',alpha=0.7)
ax.set_title("Papers on AI, ML and DL, total per year (bar) and cumulative (red)",size=14)
#What are the categories of the papers? Are we capturing what we think we are capturing
#Top 20
all_papers['category'].value_counts()[:20] | notebooks/ml_topic_analysis_exploration.ipynb | Juan-Mateos/coll_int_ai_case | mit |
See <a href='https://arxiv.org/help/api/user-manual'>here</a> for abbreviations of categories.
In a nutshell, AI is AI, LG is 'Learning', CV is 'Computer Vision', 'CL' is 'computation and language' and NE is 'Neural and Evolutionary computing'. SL.ML is kind of self-explanatory. We seem to be picking up the main things | #NB do we want to remove hyphens?
punct = re.sub('-','',st.punctuation)
def comp_sentence(sentence):
'''
Takes a sentence and pre-processes it.
The output is the sentence as a bag of words
'''
#Remove line breaks and hyphens
sentence = re.sub('\n',' ',sentence)
sentence = re.sub('-',' ',sentence)
#Lowercase and tokenise
text_lowered = [x.lower() for x in sentence.split(" ")]
#Remove signs and digits
text_no_signs_digits = ["".join([x for x in el if x not in punct+st.digits]) for
el in text_lowered]
#Remove stop words, single letters
text_stopped = [w for w in text_no_signs_digits if w not in stopwords_c and
len(w)>1]
#Stem
text_lemmatised = [lemmatizer.lemmatize(w) for w in text_stopped]
#Output
return(text_lemmatised)
#Process text
clean_corpus = [comp_sentence(x) for x in all_papers['summary']]
#We remove rate words
word_freqs = pd.Series([x for el in clean_corpus for x in el]).value_counts()
word_freqs[:30]
rare_words = word_freqs.index[word_freqs<=2]
rare_words[:10] | notebooks/ml_topic_analysis_exploration.ipynb | Juan-Mateos/coll_int_ai_case | mit |
Lots of the rare words seem to be typos and so forth. We remove them | #Removing rare words
clean_corpus_no_rare = [[x for x in el if x not in rare_words] for el in clean_corpus] | notebooks/ml_topic_analysis_exploration.ipynb | Juan-Mateos/coll_int_ai_case | mit |
2 NLP (topic modelling & word embeddings) | #Identify 2-grams (frequent in science!)
bigram_transformer = gensim.models.Phrases(clean_corpus_no_rare)
#Train the model on the corpus
#Let's do a bit of grid search
#model = gensim.models.Word2Vec(bigram_transformer[clean_corpus], size=360, window=15, min_count=2, iter=20)
model.most_similar('ai_safety')
model.most_similar('complexity')
model.most_similar('github')
#Create 3 different dictionaries and bows depending on word sizes
def remove_words_below_threshold(corpus,threshold):
'''
Takes a list of terms and removes any which are below a threshold of occurrences
'''
#Produce token frequencies
token_frequencies = pd.Series([x for el in corpus for x in el]).value_counts()
#Identify tokens to drop (below a threshold)
tokens_to_drop = token_frequencies.index[token_frequencies<=threshold]
#Processed corpus
processed_corpus = [[x for x in el if x not in tokens_to_drop] for el in corpus]
#Dictionary
dictionary = gensim.corpora.Dictionary(processed_corpus)
corpus_bow = [dictionary.doc2bow(x) for x in processed_corpus]
return([dictionary,corpus_bow,processed_corpus])
#Initial model run to see what comes out.
#Transform corpus to bigrams
transformed_corpus = bigram_transformer[clean_corpus]
corpora_to_process = {str(x):remove_words_below_threshold(transformed_corpus,x) for x in [1,2,5,10]}
#Need to turn this into a function.
#Topic modelling
#Parameters for Grid search.
lda_params = list(product([100,200,300],[2,5]))
#Model container
lda_models = []
for x in lda_params:
#Print stage
print('{x}_{y}'.format(x=x[0],y=x[1]))
#Load corpus and dict
dictionary = corpora_to_process[str(x[1])][0]
corpus_bow = corpora_to_process[str(x[1])][1]
corpus = corpora_to_process[str(x[1])][2]
print('training')
#Train model
mod = gensim.models.LdaModel(corpus_bow,num_topics=x[0],id2word=dictionary,
passes=10,iterations=50)
print('coherence')
#Extract coherence
cm = CoherenceModel(mod,texts=corpus,
dictionary=dictionary,coherence='u_mass')
#Get value
try:
coherence_value = cm.get_coherence()
except:
print('coherence_error')
coherence_value='error'
lda_models.append([x,mod,[coherence_value,cm]])
with open(mod_path+'/{t}_ai_topic_models.p'.format(t=today_str),'wb') as outfile:
pickle.dump(lda_models,outfile)
#Visualiase model performance
model_eval = pd.DataFrame([[x[0][0],x[0][1],x[2][0]] for x in lda_models],columns=['topics','word_lim','coherence'])
fig,ax = plt.subplots(figsize=(10,5))
cols = ['red','green','blue']
legs = []
for num,x in enumerate(set(model_eval['word_lim'])):
subset = model_eval.loc[[z == x for z in model_eval['word_lim']],:]
ax.plot(subset.loc[:,'topics'],subset.loc[:,'coherence'],color=cols[num-1])
legs.append([cols[num-1],x])
ax.legend(labels=[x[1] for x in legs],title='Min word count')
ax.set_title('Model performance with different parameters')
with open(mod_path+'/19_8_2017_ai_topic_models.p','rb') as infile:
lda_models = pickle.load(infile)
check_model= lda_models[1][1]
#Explore topics via LDAvis
import pyLDAvis.gensim
pyLDAvis.enable_notebook()
pyLDAvis.gensim.prepare(
#Insert best model/corpus/topics here
check_model,
corpora_to_process[str(5)][1],
corpora_to_process[str(5)][0])
#Can we extract the relevant terms for the topics as in Sievert and Shirley in order to name them?
#First - create a matrix with top 30 terms per topic
top_30_kws = [check_model.get_topic_terms(topicid=n,topn=1000) for n in np.arange(0,100)]
#Keyword df where the columns are tokens and the rows are topics
top_30_kws_df = pd.concat([pd.DataFrame([x[1] for x in el],
index=[x[0] for x in el]) for el in top_30_kws],
axis=1).fillna(0).T.reset_index(drop=True)
#This is the dictionary
selected_dictionary = corpora_to_process[str(5)][0]
#Total number of terms in the document
total_terms = np.sum([vals for vals in selected_dictionary.dfs.values()])
#Appearances of different terms
document_freqs = pd.Series([v for v in selected_dictionary.dfs.values()],
index=[k for k in selected_dictionary.dfs.keys()])[top_30_kws_df.columns]/total_terms
#Normalise the terms (divide the vector of probabilities of each keywords in each topic by the totals)
top_30_kws_normalised = top_30_kws_df.apply(lambda x: x/document_freqs,axis=0)
#Now we want to extract, for each topic, the relevance score.
def relevance_score(prob_in_topic,prob_in_corpus,id2word_lookup,lambda_par = 0.6):
'''
Combines the probabilities using the definition in Sievert and Shirley and returns the top 5 named
#terms for each topic
'''
#Create dataframe
combined = pd.concat([prob_in_topic,prob_in_corpus],axis=1)
combined.columns=['prob_in_topic','prob_in_corpus']
#Create relevance metric
combined['relevance'] = lambda_par*combined['prob_in_topic'] + (1-lambda_par)*combined['prob_in_corpus']
#Top words
top_ids = list(combined.sort_values('relevance',ascending=False).index[:5])
#Top words
top_words = "_".join([id2word_lookup[this_id] for this_id in top_ids])
return(top_words)
relevance_scores = [relevance_score(top_30_kws_df.iloc[n,:],
top_30_kws_normalised.iloc[n,:],
dictionary.id2token,lambda_par=0.6) for n in np.arange(len(top_30_kws_df))]
%%time
#Create a df with the topic predictions.
paper_preds = check_model[corpora_to_process[str(5)][1]]
paper_topics_df = pd.concat([pd.DataFrame([x[1] for x in el],index=[x[0] for x in el]) for el in paper_preds],
axis=1).T
#Replace NAs with zeros and drop pointless index
paper_topics_df.fillna(value=0,inplace=True)
paper_topics_df.reset_index(drop=True,inplace=True)
paper_topics_df.columns = relevance_scores
paper_topics_df.to_csv(int_data+'/{t}_paper_topic_mix.csv'.format(t=today_str),index=False)
#paper_topics_df = pd.read_csv(int_data+'/{t}_paper_topic_mix.csv')
#Quick test of Deep learning papers
#These are papers with a topic that seems to capture deep learning
dl_papers = [x>0.05 for x in paper_topics_df['network_training_model_deep_deep_learning']]
dl_papers_metadata = pd.concat([pd.Series(dl_papers),all_papers],axis=1)
paper_frequencies = pd.crosstab(dl_papers_metadata.year_published,dl_papers_metadata[0])
paper_frequencies.columns=['no_dl','dl']
fig,ax = plt.subplots(figsize=(10,5))
paper_frequencies.plot.bar(stacked=True,ax=ax)
ax.set_title('Number of papers in the DL \'topic\'')
ax.legend(labels=['Not ANN/DL related','NN/DL topic >0.05']) | notebooks/ml_topic_analysis_exploration.ipynb | Juan-Mateos/coll_int_ai_case | mit |
Some of this is interesting. Doesn't seem to be picking up the policy related terms (safety, discrimination)
Next stages - focus on policy related terms. Can we look for papers in keyword dictionaries identified through the word embeddings?
Obtain Google Scholar data | #How many authors are there in the data? Can we collect all their institutions from Google Scholar
paper_authors = pd.Series([x for el in all_papers['authors'] for x in el.split(", ")])
paper_authors_unique = paper_authors.drop_duplicates()
len(paper_authors_unique) | notebooks/ml_topic_analysis_exploration.ipynb | Juan-Mateos/coll_int_ai_case | mit |
We have 68,000 authors. It might take a while to get their data from Google Scholar | #Top authors and frequencies
authors_freq = paper_authors.value_counts()
fig,ax=plt.subplots(figsize=(10,3))
ax.hist(authors_freq,bins=30)
ax.set_title('Distribution of publications')
#Pretty skewed distribution!
print(authors_freq.describe())
np.sum(authors_freq>2) | notebooks/ml_topic_analysis_exploration.ipynb | Juan-Mateos/coll_int_ai_case | mit |
Less than 10,000 authors with 3+ papers in the data | get_scholar_data(
%%time
#Test run
import scholarly
@ratelim.patient(max_calls=30,time_interval=60)
def get_scholar_data(scholarly_object):
''''''
try:
scholarly_object = next(scholarly_object)
metadata = {}
metadata['name']=scholarly_object.name
metadata['affiliation'] = scholarly_object.affiliation
metadata['interests'] = scholarly_object.interests
return(metadata)
except:
return('nothing')
#Extract information from each query (it is a generator)
#Get data
#ml_author_gscholar=[]
for num,x in enumerate(paper_authors_unique[1484:]):
if num % 100 == 0:
print(str(num)+":"+x)
result = get_scholar_data(scholarly.search_author(x))
ml_author_gscholar.append(result)
len(ml_author_gscholar) | notebooks/ml_topic_analysis_exploration.ipynb | Juan-Mateos/coll_int_ai_case | mit |
1. General Mixture Models
pomegranate has a very efficient implementation of mixture models, particularly Gaussian mixture models. Lets take a look at how fast pomegranate is versus sklearn, and then see how much faster parallelization can get it to be. | n, d, k = 1000000, 5, 3
X, y = create_dataset(n, d, k)
print "sklearn GMM"
%timeit GaussianMixture(n_components=k, covariance_type='full', max_iter=15, tol=1e-10).fit(X)
print
print "pomegranate GMM"
%timeit GeneralMixtureModel.from_samples(MultivariateGaussianDistribution, k, X, max_iterations=15, stop_threshold=1e-10)
print
print "pomegranate GMM (4 jobs)"
%timeit GeneralMixtureModel.from_samples(MultivariateGaussianDistribution, k, X, n_jobs=4, max_iterations=15, stop_threshold=1e-10) | tutorials/old/Tutorial_7_Parallelization.ipynb | jmschrei/pomegranate | mit |
It looks like on a large dataset not only is pomegranate faster than sklearn at performing 15 iterations of EM on 3 million 5 dimensional datapoints with 3 clusters, but the parallelization is able to help in speeding things up.
Lets now take a look at the time it takes to make predictions using GMMs. Lets fit the model to a small amount of data, and then predict a larger amount of data drawn from the same underlying distributions. | d, k = 25, 2
X, y = create_dataset(1000, d, k)
a = GaussianMixture(k, n_init=1, max_iter=25).fit(X)
b = GeneralMixtureModel.from_samples(MultivariateGaussianDistribution, k, X, max_iterations=25)
del X, y
n = 1000000
X, y = create_dataset(n, d, k)
print "sklearn GMM"
%timeit -n 1 a.predict_proba(X)
print
print "pomegranate GMM"
%timeit -n 1 b.predict_proba(X)
print
print "pomegranate GMM (4 jobs)"
%timeit -n 1 b.predict_proba(X, n_jobs=4) | tutorials/old/Tutorial_7_Parallelization.ipynb | jmschrei/pomegranate | mit |
It looks like pomegranate can be slightly slower than sklearn when using a single processor, but that it can be parallelized to get faster performance. At the same time, predictions at this level happen so quickly (millions per second) that this may not be the most reliable test for parallelization.
To ensure that we're getting the exact same results just faster, lets subtract the predictions from each other and make sure that the sum is equal to 0. | print (b.predict_proba(X) - b.predict_proba(X, n_jobs=4)).sum() | tutorials/old/Tutorial_7_Parallelization.ipynb | jmschrei/pomegranate | mit |
Great, no difference between the two.
Lets now make sure that pomegranate and sklearn are learning basically the same thing. Lets fit both models to some 2 dimensional 2 component data and make sure that they both extract the underlying clusters by plotting them. | d, k = 2, 2
X, y = create_dataset(1000, d, k, alpha=2)
a = GaussianMixture(k, n_init=1, max_iter=25).fit(X)
b = GeneralMixtureModel.from_samples(MultivariateGaussianDistribution, k, X, max_iterations=25)
y1, y2 = a.predict(X), b.predict(X)
plt.figure(figsize=(16,6))
plt.subplot(121)
plt.title("sklearn clusters", fontsize=14)
plt.scatter(X[y1==0, 0], X[y1==0, 1], color='m', edgecolor='m')
plt.scatter(X[y1==1, 0], X[y1==1, 1], color='c', edgecolor='c')
plt.subplot(122)
plt.title("pomegranate clusters", fontsize=14)
plt.scatter(X[y2==0, 0], X[y2==0, 1], color='m', edgecolor='m')
plt.scatter(X[y2==1, 0], X[y2==1, 1], color='c', edgecolor='c') | tutorials/old/Tutorial_7_Parallelization.ipynb | jmschrei/pomegranate | mit |
It looks like we're getting the same basic results for the two. The two algorithms are initialized a bit differently, and so it can be difficult to directly compare the results between them, but it looks like they're getting roughly the same results.
3. Multivariate Gaussian HMM
Now let's move on to training a hidden Markov model with multivariate Gaussian emissions with a diagonal covariance matrix. We'll randomly generate some Gaussian distributed numbers and use pomegranate with either one or four threads to fit our model to the data. | X = numpy.random.randn(1000, 500, 50)
print "pomegranate Gaussian HMM (1 job)"
%timeit -n 1 -r 1 HiddenMarkovModel.from_samples(NormalDistribution, 5, X, max_iterations=5)
print
print "pomegranate Gaussian HMM (2 jobs)"
%timeit -n 1 -r 1 HiddenMarkovModel.from_samples(NormalDistribution, 5, X, max_iterations=5, n_jobs=2)
print
print "pomegranate Gaussian HMM (2 jobs)"
%timeit -n 1 -r 1 HiddenMarkovModel.from_samples(NormalDistribution, 5, X, max_iterations=5, n_jobs=4) | tutorials/old/Tutorial_7_Parallelization.ipynb | jmschrei/pomegranate | mit |
All we had to do was pass in the n_jobs parameter to the fit function in order to get a speed improvement. It looks like we're getting a really good speed improvement, as well! This is mostly because the HMM algorithms perform a lot more operations than the other models, and so spend the vast majority of time with the GIL released. You may not notice as strong speedups when using a MultivariateGaussianDistribution because BLAS uses multithreaded operations already internally, even when only one job is specified.
Now lets look at the prediction function to make sure the we're getting speedups there as well. You'll have to use a wrapper function to parallelize the predictions for a HMM because it returns an annotated sequence rather than a single value like a classic machine learning model might. | model = HiddenMarkovModel.from_samples(NormalDistribution, 5, X, max_iterations=2, verbose=False)
print "pomegranate Gaussian HMM (1 job)"
%timeit predict_proba(model, X)
print
print "pomegranate Gaussian HMM (2 jobs)"
%timeit predict_proba(model, X, n_jobs=2) | tutorials/old/Tutorial_7_Parallelization.ipynb | jmschrei/pomegranate | mit |
Great, we're getting a really good speedup on that as well! Looks like the parallel processing is more efficient with a bigger, more complex model, than with a simple one. This can make sense, because all inference/training is more complex, and so there is more time with the GIL released compared to with the simpler operations.
4. Mixture of Hidden Markov Models
Let's stack another layer onto this model by making it a mixture of these hidden Markov models, instead of a single one. At this point we're sticking a multivariate Gaussian HMM into a mixture and we're going to train this big thing in parallel. | def create_model(mus):
n = mus.shape[0]
starts = numpy.zeros(n)
starts[0] = 1.
ends = numpy.zeros(n)
ends[-1] = 0.5
transition_matrix = numpy.zeros((n, n))
distributions = []
for i in range(n):
transition_matrix[i, i] = 0.5
if i < n - 1:
transition_matrix[i, i+1] = 0.5
distribution = IndependentComponentsDistribution([NormalDistribution(mu, 1) for mu in mus[i]])
distributions.append(distribution)
model = HiddenMarkovModel.from_matrix(transition_matrix, distributions, starts, ends)
return model
def create_mixture(mus):
hmms = [create_model(mu) for mu in mus]
return GeneralMixtureModel(hmms)
n, d = 50, 10
mus = [(numpy.random.randn(d, n)*0.2 + numpy.random.randn(n)*2).T for i in range(2)]
model = create_mixture(mus)
X = numpy.random.randn(400, 150, d)
print "pomegranate Mixture of Gaussian HMMs (1 job)"
%timeit model.fit(X, max_iterations=5)
print
model = create_mixture(mus)
print "pomegranate Mixture of Gaussian HMMs (2 jobs)"
%timeit model.fit(X, max_iterations=5, n_jobs=2) | tutorials/old/Tutorial_7_Parallelization.ipynb | jmschrei/pomegranate | mit |
Looks like we're getting a really nice speed improvement when training this complex model. Let's take a look now at the time it takes to do inference with it. | model = create_mixture(mus)
print "pomegranate Mixture of Gaussian HMMs (1 job)"
%timeit model.predict_proba(X)
print
model = create_mixture(mus)
print "pomegranate Mixture of Gaussian HMMs (2 jobs)"
%timeit model.predict_proba(X, n_jobs=2) | tutorials/old/Tutorial_7_Parallelization.ipynb | jmschrei/pomegranate | mit |
The inner product of blades in GAlgebra is zero if either operand is a scalar:
$$\begin{split}\begin{aligned}
{\boldsymbol{A}}{r}{\wedge}{\boldsymbol{B}}{s} &\equiv {\left <{{\boldsymbol{A}}{r}{\boldsymbol{B}}{s}} \right >{r+s}} \
{\boldsymbol{A}}{r}\cdot{\boldsymbol{B}}{s} &\equiv {\left { { \begin{array}{cc}
r\mbox{ and }s \ne 0: & {\left <{{\boldsymbol{A}}{r}{\boldsymbol{B}}{s}} \right >{{\left |{r-s}\right |}}} \
r\mbox{ or }s = 0: & 0 \end{array}} \right }}
\end{aligned}\end{split}$$
This definition comes from David Hestenes and Garret Sobczyk, “Clifford Algebra to Geometric Calculus,” Kluwer Academic Publishers, 1984.
In some other literature, the inner product is defined without the exceptional case for scalar part and the definition above is known as "the modified Hestenes inner product" (this name comes from the source code of GAViewer). | c|a
a|c
c|A
A|c | examples/ipython/inner_product.ipynb | arsenovic/galgebra | bsd-3-clause |
$ab=a \wedge b + a \cdot b$ holds for vectors: | a*b
a^b
a|b
(a*b)-(a^b)-(a|b) | examples/ipython/inner_product.ipynb | arsenovic/galgebra | bsd-3-clause |
$aA=a \wedge A + a \cdot A$ holds for the products between vectors and multivectors: | a*A
a^A
a|A
(a*A)-(a^A)-(a|A) | examples/ipython/inner_product.ipynb | arsenovic/galgebra | bsd-3-clause |
$AB=A \wedge B + A \cdot B$ does NOT hold for the products between multivectors and multivectors: | A*B
A|B
(A*B)-(A^B)-(A|B)
(A<B)+(A|B)+(A>B)-A*B | examples/ipython/inner_product.ipynb | arsenovic/galgebra | bsd-3-clause |
Toolkit: Visualization Functions
This class will introduce 3 different visualizations that can be used with the two different classification type neural networks and regression neural networks.
Confusion Matrix - For any type of classification neural network.
ROC Curve - For binary classification.
Lift Curve - For regression neural networks.
The code used to produce these visualizations is shown here: | %matplotlib inline
import matplotlib.pyplot as plt
from sklearn.metrics import roc_curve, auc
# Plot a confusion matrix.
# cm is the confusion matrix, names are the names of the classes.
def plot_confusion_matrix(cm, names, title='Confusion matrix', cmap=plt.cm.Blues):
plt.imshow(cm, interpolation='nearest', cmap=cmap)
plt.title(title)
plt.colorbar()
tick_marks = np.arange(len(names))
plt.xticks(tick_marks, names, rotation=45)
plt.yticks(tick_marks, names)
plt.tight_layout()
plt.ylabel('True label')
plt.xlabel('Predicted label')
# Plot an ROC. pred - the predictions, y - the expected output.
def plot_roc(pred,y):
fpr, tpr, _ = roc_curve(y_test, pred)
roc_auc = auc(fpr, tpr)
plt.figure()
plt.plot(fpr, tpr, label='ROC curve (area = %0.2f)' % roc_auc)
plt.plot([0, 1], [0, 1], 'k--')
plt.xlim([0.0, 1.0])
plt.ylim([0.0, 1.05])
plt.xlabel('False Positive Rate')
plt.ylabel('True Positive Rate')
plt.title('Receiver Operating Characteristic (ROC)')
plt.legend(loc="lower right")
plt.show()
# Plot a lift curve. pred - the predictions, y - the expected output.
def chart_regression(pred,y):
t = pd.DataFrame({'pred' : pred.flatten(), 'y' : y_test.flatten()})
t.sort_values(by=['y'],inplace=True)
a = plt.plot(t['y'].tolist(),label='expected')
b = plt.plot(t['pred'].tolist(),label='prediction')
plt.ylabel('output')
plt.legend()
plt.show() | t81_558_class4_class_reg.ipynb | jbliss1234/ML | apache-2.0 |
Binary Classification
Binary classification is used to create a model that classifies between only two classes. These two classes are often called "positive" and "negative". Consider the following program that uses the wcbreast_wdbc dataset to classify if a breast tumor is cancerous (malignant) or not (benign). The iris dataset is not binary, because there are three classes (3 types of iris). | import os
import pandas as pd
from sklearn.cross_validation import train_test_split
import tensorflow.contrib.learn as skflow
import numpy as np
from sklearn import metrics
path = "./data/"
filename = os.path.join(path,"wcbreast_wdbc.csv")
df = pd.read_csv(filename,na_values=['NA','?'])
# Encode feature vector
df.drop('id',axis=1,inplace=True)
encode_numeric_zscore(df,'mean_radius')
encode_text_index(df,'mean_texture')
encode_text_index(df,'mean_perimeter')
encode_text_index(df,'mean_area')
encode_text_index(df,'mean_smoothness')
encode_text_index(df,'mean_compactness')
encode_text_index(df,'mean_concavity')
encode_text_index(df,'mean_concave_points')
encode_text_index(df,'mean_symmetry')
encode_text_index(df,'mean_fractal_dimension')
encode_text_index(df,'se_radius')
encode_text_index(df,'se_texture')
encode_text_index(df,'se_perimeter')
encode_text_index(df,'se_area')
encode_text_index(df,'se_smoothness')
encode_text_index(df,'se_compactness')
encode_text_index(df,'se_concavity')
encode_text_index(df,'se_concave_points')
encode_text_index(df,'se_symmetry')
encode_text_index(df,'se_fractal_dimension')
encode_text_index(df,'worst_radius')
encode_text_index(df,'worst_texture')
encode_text_index(df,'worst_perimeter')
encode_text_index(df,'worst_area')
encode_text_index(df,'worst_smoothness')
encode_text_index(df,'worst_compactness')
encode_text_index(df,'worst_concavity')
encode_text_index(df,'worst_concave_points')
encode_text_index(df,'worst_symmetry')
encode_text_index(df,'worst_fractal_dimension')
diagnosis = encode_text_index(df,'diagnosis')
num_classes = len(diagnosis)
# Create x & y for training
# Create the x-side (feature vectors) of the training
x, y = to_xy(df,'diagnosis')
# Split into train/test
x_train, x_test, y_train, y_test = train_test_split(
x, y, test_size=0.25, random_state=42)
# Create a deep neural network with 3 hidden layers of 10, 20, 10
classifier = skflow.TensorFlowDNNClassifier(hidden_units=[10, 20, 10], n_classes=num_classes,
steps=10000)
# Early stopping
early_stop = skflow.monitors.ValidationMonitor(x_test, y_test,
early_stopping_rounds=200, print_steps=50, n_classes=num_classes)
# Fit/train neural network
classifier.fit(x_train, y_train, early_stop)
# Measure accuracy
score = metrics.accuracy_score(y, classifier.predict(x))
print("Final accuracy: {}".format(score))
| t81_558_class4_class_reg.ipynb | jbliss1234/ML | apache-2.0 |
Confusion Matrix
The confusion matrix is a common visualization for both binary and larger classification problems. Often a model will have difficulty differentiating between two classes. For example, a neural network might be really good at telling the difference between cats and dogs, but not so good at telling the difference between dogs and wolves. The following code generates a confusion matrix: | import numpy as np
from sklearn import svm, datasets
from sklearn.cross_validation import train_test_split
from sklearn.metrics import confusion_matrix
pred = classifier.predict(x_test)
# Compute confusion matrix
cm = confusion_matrix(y_test, pred)
np.set_printoptions(precision=2)
print('Confusion matrix, without normalization')
print(cm)
plt.figure()
plot_confusion_matrix(cm, diagnosis)
# Normalize the confusion matrix by row (i.e by the number of samples
# in each class)
cm_normalized = cm.astype('float') / cm.sum(axis=1)[:, np.newaxis]
print('Normalized confusion matrix')
print(cm_normalized)
plt.figure()
plot_confusion_matrix(cm_normalized, diagnosis, title='Normalized confusion matrix')
plt.show() | t81_558_class4_class_reg.ipynb | jbliss1234/ML | apache-2.0 |
The above two confusion matrixes show the same network. The bottom (normalized) is the type you will normally see. Notice the two labels. The label "B" means benign (no cancer) and the label "M" means malignant (cancer). The left-right (x) axis are the predictions, the top-bottom) are the expected outcomes. A perfect model (that never makes an error) has a dark blue diagonal that runs from top-left to bottom-right.
To read, consider the top-left square. This square indicates "true labeled" of B and also "predicted label" of B. This is good! The prediction matched the truth. The blueness of this box represents how often "B" is classified correct. It is not darkest blue. This is because the square to the right(which is off the perfect diagonal) has some color. This square indicates truth of "B" but prediction of "M". The white square, at the bottom-left, indicates a true of "M" but predicted of "B". The whiteness indicates this rarely happens.
Your conclusion from the above chart is that the model sometimes classifies "B" as "M" (a false negative), but never mis-classifis "M" as "B". Always look for the dark diagonal, this is good!
ROC Curves
ROC curves can be a bit confusing. However, they are very common. It is important to know how to read them. Even their name is confusing. Do not worry about their name, it comes from electrical engineering (EE).
Binary classification is common in medical testing. Often you want to diagnose if someone has a disease. This can lead to two types of errors, know as false positives and false negatives:
False Positive - Your test (neural network) indicated that the patient had the disease; however, the patient did not have the disease.
False Negative - Your test (neural network) indicated that the patient did not have the disease; however, the patient did have the disease.
True Positive - Your test (neural network) correctly identified that the patient had the disease.
True Negative - Your test (neural network) correctly identified that the patient did not have the disease.
Types of errors:
Neural networks classify in terms of probbility of it being positive. However, at what probability do you give a positive result? Is the cutoff 50%? 90%? Where you set this cutoff is called the threshold. Anything above the cutoff is positive, anything below is negative. Setting this cutoff allows the model to be more sensative or specific:
The following shows a more sensitive cutoff:
An ROC curve measures how good a model is regardless of the cutoff. The following shows how to read a ROC chart:
The following code shows an ROC chart for the breast cancer neural network. The area under the curve (AUC) is also an important measure. The larger the AUC, the better. | pred = classifier.predict_proba(x_test)
pred = pred[:,1] # Only positive cases
# print(pred[:,1])
plot_roc(pred,y_test)
| t81_558_class4_class_reg.ipynb | jbliss1234/ML | apache-2.0 |
Classification
We've already seen multi-class classification, with the iris dataset. Confusion matrixes work just fine with 3 classes. The following code generates a confusion matrix for iris. | import os
import pandas as pd
from sklearn.cross_validation import train_test_split
import tensorflow.contrib.learn as skflow
import numpy as np
path = "./data/"
filename = os.path.join(path,"iris.csv")
df = pd.read_csv(filename,na_values=['NA','?'])
# Encode feature vector
encode_numeric_zscore(df,'petal_w')
encode_numeric_zscore(df,'petal_l')
encode_numeric_zscore(df,'sepal_w')
encode_numeric_zscore(df,'sepal_l')
species = encode_text_index(df,"species")
num_classes = len(species)
# Create x & y for training
# Create the x-side (feature vectors) of the training
x, y = to_xy(df,'species')
# Split into train/test
x_train, x_test, y_train, y_test = train_test_split(
x, y, test_size=0.25, random_state=45)
# as much as I would like to use 42, it gives a perfect result, and a boring confusion matrix!
# Create a deep neural network with 3 hidden layers of 10, 20, 10
classifier = skflow.TensorFlowDNNClassifier(hidden_units=[10, 20, 10], n_classes=num_classes,
steps=10000)
# Early stopping
early_stop = skflow.monitors.ValidationMonitor(x_test, y_test,
early_stopping_rounds=200, print_steps=50, n_classes=num_classes)
# Fit/train neural network
classifier.fit(x_train, y_train, early_stop)
import numpy as np
from sklearn import svm, datasets
from sklearn.cross_validation import train_test_split
from sklearn.metrics import confusion_matrix
pred = classifier.predict(x_test)
# Compute confusion matrix
cm = confusion_matrix(y_test, pred)
np.set_printoptions(precision=2)
print('Confusion matrix, without normalization')
print(cm)
plt.figure()
plot_confusion_matrix(cm, species)
# Normalize the confusion matrix by row (i.e by the number of samples
# in each class)
cm_normalized = cm.astype('float') / cm.sum(axis=1)[:, np.newaxis]
print('Normalized confusion matrix')
print(cm_normalized)
plt.figure()
plot_confusion_matrix(cm_normalized, species, title='Normalized confusion matrix')
plt.show() | t81_558_class4_class_reg.ipynb | jbliss1234/ML | apache-2.0 |
See the strong diagonal? Iris is easy. See the light blue near the bottom? Sometimes virginica is confused for versicolor.
Regression
We've already seen regression with the MPG dataset. Regression uses its own set of visualizations, one of the most common is the lift chart. The following code generates a lift chart. | import tensorflow.contrib.learn as skflow
import pandas as pd
import os
import numpy as np
from sklearn import metrics
from scipy.stats import zscore
path = "./data/"
filename_read = os.path.join(path,"auto-mpg.csv")
df = pd.read_csv(filename_read,na_values=['NA','?'])
# create feature vector
missing_median(df, 'horsepower')
df.drop('name',1,inplace=True)
encode_numeric_zscore(df, 'horsepower')
encode_numeric_zscore(df, 'weight')
encode_numeric_zscore(df, 'cylinders')
encode_numeric_zscore(df, 'displacement')
encode_numeric_zscore(df, 'acceleration')
encode_text_dummy(df, 'origin')
# Encode to a 2D matrix for training
x,y = to_xy(df,['mpg'])
# Split into train/test
x_train, x_test, y_train, y_test = train_test_split(
x, y, test_size=0.25, random_state=42)
# Create a deep neural network with 3 hidden layers of 50, 25, 10
regressor = skflow.TensorFlowDNNRegressor(hidden_units=[50, 25, 10], steps=5000)
# Early stopping
early_stop = skflow.monitors.ValidationMonitor(x_test, y_test,
early_stopping_rounds=200, print_steps=50)
# Fit/train neural network
regressor.fit(x_train, y_train, early_stop)
pred = regressor.predict(x_test)
chart_regression(pred,y_test) | t81_558_class4_class_reg.ipynb | jbliss1234/ML | apache-2.0 |
Reordering the Callendar-Van Duzen equation we obtain the following
$$ AT+BT^2+C(T-100)T^3 =\frac{R(T)}{R_0}-1 \enspace,$$
which we can write in matrix form as $Mx=p$, where
$$\begin{bmatrix} T_1 & T_1^2 & (T_1-100)T_1^3 \ T_2 & T_2^2 & (T_2-100)T_2^3 \ T_3 & T_3^2 & (T_3-100)T_3^3\end{bmatrix} \begin{bmatrix} A\ B \ C\end{bmatrix} = \begin{bmatrix} \frac{R(T_1)}{R_0}-1 \ \frac{R(T_2)}{R_0}-1 \ \frac{R(T_3)}{R_0}-1\end{bmatrix} \enspace.$$
Because $M$ is square we can solve by computing $M^{-1}$ directly. | R0=25;
M=np.array([[T_exp[0],(T_exp[0])**2,(T_exp[0]-100)*(T_exp[0])**3],[T_exp[1],(T_exp[1])**2,(T_exp[1]-100)*(T_exp[1])**3],[T_exp[2],(T_exp[2])**2,(T_exp[2]-100)*(T_exp[2])**3]]);
p=np.array([[(R_exp[0]/R0)-1],[(R_exp[1]/R0)-1],[(R_exp[2]/R0)-1]]);
x = np.linalg.solve(M,p) #solve linear equations system
np.set_printoptions(precision=3)
print('M')
print(M)
print('\n')
print('p')
print(p)
print('\n')
print('x')
print(x) | notebooks/Ex_2_3.ipynb | agmarrugo/sensors-actuators | mit |
We have found the coeffiecients $A$, $B$, and $C$ necessary to describe the sensor's transfer function. Now we plot it from -200 C a 600 C. | A=x[0];B=x[1];C=x[2];
T_range= np.arange(start = -200, stop = 601, step = 1);
R_funT= R0*(1+A[0]*T_range+B[0]*(T_range)**2+C[0]*(T_range-100)*(T_range)**3);
plt.plot(T_range,R_funT,T_exp[0],R_exp[0],'ro',T_exp[1],R_exp[1],'ro',T_exp[2],R_exp[2],'ro');
plt.ylabel('Sensor resistance [Ohm]')
plt.xlabel('Temperature [C]')
plt.show()
| notebooks/Ex_2_3.ipynb | agmarrugo/sensors-actuators | mit |
Reddy Mikks model
Given the following variables:
$\begin{aligned}
x_1 = \textrm{Tons of exterior paint produced daily} \newline
x_2 = \textrm{Tons of interior paint produced daily}
\end{aligned}$
and knowing that we want to maximize the profit, where \$5000 is the profit from exterior paint and \$4000 is the profit from a ton of interior paint, the Reddy Mikks model is:
$$\textrm{Maximize } z = 5x_1 + 4x_2$$
subject to
$$6x_1 + 4x_2 \le 24$$
$$x_1 + 2x_2 \le 6$$
$$-x_1 + x_2 \le 1$$
$$x_2 \le 2$$
$$x_1, x_2 \ge 0$$ | reddymikks = pywraplp.Solver('Reddy_Mikks', pywraplp.Solver.GLOP_LINEAR_PROGRAMMING)
x1 = reddymikks.NumVar(0, reddymikks.infinity(), 'x1')
x2 = reddymikks.NumVar(0, reddymikks.infinity(), 'x2')
reddymikks.Add(6*x1 + 4*x2 <= 24)
reddymikks.Add(x1 + 2*x2 <= 6)
reddymikks.Add(-x1 + x2 <= 1)
reddymikks.Add(x2 <= 2)
profit = reddymikks.Objective()
profit.SetCoefficient(x1, 5)
profit.SetCoefficient(x2, 4)
profit.SetMaximization()
status = reddymikks.Solve()
if status not in [reddymikks.OPTIMAL, reddymikks.FEASIBLE]:
raise Exception('No feasible solution found')
print("The company should produce",round(x1.solution_value(),2),"tons of exterior paint")
print("The company should produce",round(x2.solution_value(),2),"tons of interior paint")
print("The optimal profit is", profit.Value(), 'thousand USD') | Linear Programming with OR-Tools.ipynb | rayjustinhuang/DataAnalysisandMachineLearning | mit |
More simple problems
A company that operates 10 hours a day manufactures two products on three sequential processes. The following data characterizes the problem: | import pandas as pd
problemdata = pd.DataFrame({'Process 1': [10, 5], 'Process 2':[6, 20], 'Process 3':[8, 10], 'Unit profit':[20, 30]})
problemdata.index = ['Product 1', 'Product 2']
problemdata | Linear Programming with OR-Tools.ipynb | rayjustinhuang/DataAnalysisandMachineLearning | mit |
Where there are 10 hours a day dedicated to production. Process times are given in minutes per unit while profit is given in USD.
The optimal mix of the two products would be characterized by the following model:
$\begin{aligned}
x_1 = \textrm{Units of product 1} \newline
x_2 = \textrm{Units of product 2}
\end{aligned}$
$$\textrm{Maximize } z = 20x_1 + 30x_2$$
subject to
$$\begin{array}{rcl}
10x_1 + 5x_2 \le 600 \newline
6x_1 + 20x_2 \le 600 \newline
8x_1 + 10x_2 \le 600 \newline
x_1, x_2 \ge 0
\end{array}$$
(we will assume that continuous solution values are acceptable for this problem) | simpleprod = pywraplp.Solver('Simple_Production', pywraplp.Solver.GLOP_LINEAR_PROGRAMMING)
x1 = simpleprod.NumVar(0, simpleprod.infinity(), 'x1')
x2 = simpleprod.NumVar(0, simpleprod.infinity(), 'x2')
for i in problemdata.columns[:-1]:
simpleprod.Add(problemdata.loc[problemdata.index[0], i]*x1 + problemdata.loc[problemdata.index[1], i]*x2 <= 600)
profit = simpleprod.Objective()
profit.SetCoefficient(x1, 20)
profit.SetCoefficient(x2, 30)
profit.SetMaximization()
status = simpleprod.Solve()
if status not in [simpleprod.OPTIMAL, simpleprod.FEASIBLE]:
raise Exception('No feasible solution found')
print("The company should produce",round(x1.solution_value(),2),"units of product 1")
print("The company should produce",round(x2.solution_value(),2),"units of product 2")
print("The optimal profit is", round(profit.Value(),2), 'USD') | Linear Programming with OR-Tools.ipynb | rayjustinhuang/DataAnalysisandMachineLearning | mit |
1. Download Text8 Corpus | import os.path
if not os.path.isfile('text8'):
!wget -c http://mattmahoney.net/dc/text8.zip
!unzip text8.zip | docs/notebooks/nmslibtutorial.ipynb | RaRe-Technologies/gensim | lgpl-2.1 |
Import & Set up Logging
I'm not going to set up logging due to the verbose input displaying in notebooks, but if you want that, uncomment the lines in the cell below. | LOGS = False
if LOGS:
import logging
logging.basicConfig(format='%(asctime)s : %(levelname)s : %(message)s', level=logging.INFO) | docs/notebooks/nmslibtutorial.ipynb | RaRe-Technologies/gensim | lgpl-2.1 |
2. Build Word2Vec Model | from gensim.models import Word2Vec, KeyedVectors
from gensim.models.word2vec import Text8Corpus
# Using params from Word2Vec_FastText_Comparison
params = {
'alpha': 0.05,
'size': 100,
'window': 5,
'iter': 5,
'min_count': 5,
'sample': 1e-4,
'sg': 1,
'hs': 0,
'negative': 5
}
model = Word2Vec(Text8Corpus('text8'), **params)
print(model) | docs/notebooks/nmslibtutorial.ipynb | RaRe-Technologies/gensim | lgpl-2.1 |
See the Word2Vec tutorial for how to initialize and save this model.
Comparing the traditional implementation, Annoy and Nmslib approximation | # Set up the model and vector that we are using in the comparison
from gensim.similarities.index import AnnoyIndexer
from gensim.similarities.nmslib import NmslibIndexer
model.init_sims()
annoy_index = AnnoyIndexer(model, 300)
nmslib_index = NmslibIndexer(model, {'M': 100, 'indexThreadQty': 1, 'efConstruction': 100}, {'efSearch': 10})
# Dry run to make sure both indices are fully in RAM
vector = model.wv.syn0norm[0]
print(model.most_similar([vector], topn=5, indexer=annoy_index))
print(model.most_similar([vector], topn=5, indexer=nmslib_index))
print(model.most_similar([vector], topn=5))
import time
import numpy as np
def avg_query_time(annoy_index=None, queries=1000):
"""
Average query time of a most_similar method over 1000 random queries,
uses annoy if given an indexer
"""
total_time = 0
for _ in range(queries):
rand_vec = model.wv.syn0norm[np.random.randint(0, len(model.wv.vocab))]
start_time = time.clock()
model.most_similar([rand_vec], topn=5, indexer=annoy_index)
total_time += time.clock() - start_time
return total_time / queries
queries = 10000
gensim_time = avg_query_time(queries=queries)
annoy_time = avg_query_time(annoy_index, queries=queries)
nmslib_time = avg_query_time(nmslib_index, queries=queries)
print("Gensim (s/query):\t{0:.5f}".format(gensim_time))
print("Annoy (s/query):\t{0:.5f}".format(annoy_time))
print("Nmslib (s/query):\t{0:.5f}".format(nmslib_time))
speed_improvement_gensim = gensim_time / nmslib_time
speed_improvement_annoy = annoy_time / nmslib_time
print ("\nNmslib is {0:.2f} times faster on average on this particular run".format(speed_improvement_gensim))
print ("\nNmslib is {0:.2f} times faster on average than annoy on this particular run".format(speed_improvement_annoy))
| docs/notebooks/nmslibtutorial.ipynb | RaRe-Technologies/gensim | lgpl-2.1 |
3. Construct Nmslib Index with model & make a similarity query
Creating an indexer
An instance of NmslibIndexer needs to be created in order to use Nmslib in gensim. The NmslibIndexer class is located in gensim.similarities.nmslib
NmslibIndexer() takes three parameters:
model: A Word2Vec or Doc2Vec model
index_params: Parameters for building nmslib indexer. index_params effects the build time and the index size. The example is {'M': 100, 'indexThreadQty': 1, 'efConstruction': 100}. Increasing the value of M and efConstruction improves the accuracy of search. However this also leads to longer indexing times. indexThreadQty is the number of thread.
query_time_params: Parameters for querying on nmslib indexer. query_time_params effects the query time and the search accuracy. The example is {'efSearch': 100}. A larger efSearch will give more accurate results, but larger query time.
More information can be found here. The relationship between parameters, build/query time, and accuracy will be investigated later in the tutorial.
Now that we are ready to make a query, lets find the top 5 most similar words to "science" in the Text8 corpus. To make a similarity query we call Word2Vec.most_similar like we would traditionally, but with an added parameter, indexer. The only supported indexerers in gensim as of now are Annoy and Nmslib. | # Building nmslib indexer
nmslib_index = NmslibIndexer(model, {'M': 100, 'indexThreadQty': 1, 'efConstruction': 100}, {'efSearch': 10})
# Derive the vector for the word "science" in our model
vector = model["science"]
# The instance of AnnoyIndexer we just created is passed
approximate_neighbors = model.most_similar([vector], topn=11, indexer=nmslib_index)
# Neatly print the approximate_neighbors and their corresponding cosine similarity values
print("Approximate Neighbors")
for neighbor in approximate_neighbors:
print(neighbor)
normal_neighbors = model.most_similar([vector], topn=11)
print("\nNormal (not nmslib-indexed) Neighbors")
for neighbor in normal_neighbors:
print(neighbor) | docs/notebooks/nmslibtutorial.ipynb | RaRe-Technologies/gensim | lgpl-2.1 |
Analyzing the results
The closer the cosine similarity of a vector is to 1, the more similar that word is to our query, which was the vector for "science". In this case the results are almostly same.
4. Verify & Evaluate performance
Persisting Indexes
You can save and load your indexes from/to disk to prevent having to construct them each time. This will create two files on disk, fname and fname.d. Both files are needed to correctly restore all attributes. | import os
fname = '/tmp/mymodel.index'
# Persist index to disk
nmslib_index.save(fname)
# Load index back
if os.path.exists(fname):
nmslib_index2 = NmslibIndexer.load(fname)
nmslib_index2.model = model
# Results should be identical to above
vector = model["science"]
approximate_neighbors2 = model.most_similar([vector], topn=11, indexer=nmslib_index2)
for neighbor in approximate_neighbors2:
print(neighbor)
assert approximate_neighbors == approximate_neighbors2 | docs/notebooks/nmslibtutorial.ipynb | RaRe-Technologies/gensim | lgpl-2.1 |
Be sure to use the same model at load that was used originally, otherwise you will get unexpected behaviors.
Save memory by memory-mapping indices saved to disk
Nmslib library has a useful feature that indices can be memory-mapped from disk. It saves memory when the same index is used by several processes.
Below are two snippets of code. First one has a separate index for each process. The second snipped shares the index between two processes via memory-mapping. The second example uses less total RAM as it is shared. | # Remove verbosity from code below (if logging active)
if LOGS:
logging.disable(logging.CRITICAL)
from multiprocessing import Process
import psutil | docs/notebooks/nmslibtutorial.ipynb | RaRe-Technologies/gensim | lgpl-2.1 |
Bad Example: Two processes load the Word2vec model from disk and create there own Nmslib indices from that model. | %%time
model.save('/tmp/mymodel.pkl')
def f(process_id):
print('Process Id: {}'.format(os.getpid()))
process = psutil.Process(os.getpid())
new_model = Word2Vec.load('/tmp/mymodel.pkl')
vector = new_model["science"]
nmslib_index = NmslibIndexer(new_model, {'M': 100, 'indexThreadQty': 1, 'efConstruction': 100}, {'efSearch': 10})
approximate_neighbors = new_model.most_similar([vector], topn=5, indexer=nmslib_index)
print('\nMemory used by process {}: {}\n---'.format(os.getpid(), process.memory_info()))
# Creating and running two parallel process to share the same index file.
p1 = Process(target=f, args=('1',))
p1.start()
p1.join()
p2 = Process(target=f, args=('2',))
p2.start()
p2.join() | docs/notebooks/nmslibtutorial.ipynb | RaRe-Technologies/gensim | lgpl-2.1 |
Good example. Two processes load both the Word2vec model and index from disk and memory-map the index | %%time
model.save('/tmp/mymodel.pkl')
def f(process_id):
print('Process Id: {}'.format(os.getpid()))
process = psutil.Process(os.getpid())
new_model = Word2Vec.load('/tmp/mymodel.pkl')
vector = new_model["science"]
nmslib_index = NmslibIndexer.load('/tmp/mymodel.index')
nmslib_index.model = new_model
approximate_neighbors = new_model.most_similar([vector], topn=5, indexer=nmslib_index)
print('\nMemory used by process {}: {}\n---'.format(os.getpid(), process.memory_info()))
# Creating and running two parallel process to share the same index file.
p1 = Process(target=f, args=('1',))
p1.start()
p1.join()
p2 = Process(target=f, args=('2',))
p2.start()
p2.join() | docs/notebooks/nmslibtutorial.ipynb | RaRe-Technologies/gensim | lgpl-2.1 |
5. Evaluate relationship of parameters to initialization/query time and accuracy, compared with annoy | import matplotlib.pyplot as plt
%matplotlib inline | docs/notebooks/nmslibtutorial.ipynb | RaRe-Technologies/gensim | lgpl-2.1 |
Build dataset of Initialization times and accuracy measures | exact_results = [element[0] for element in model.most_similar([model.wv.syn0norm[0]], topn=100)]
# For calculating query time
queries = 1000
def create_evaluation_graph(x_values, y_values_init, y_values_accuracy, y_values_query, param_name):
plt.figure(1, figsize=(12, 6))
plt.subplot(231)
plt.plot(x_values, y_values_init)
plt.title("{} vs initalization time".format(param_name))
plt.ylabel("Initialization time (s)")
plt.xlabel(param_name)
plt.subplot(232)
plt.plot(x_values, y_values_accuracy)
plt.title("{} vs accuracy".format(param_name))
plt.ylabel("% accuracy")
plt.xlabel(param_name)
plt.tight_layout()
plt.subplot(233)
plt.plot(y_values_init, y_values_accuracy)
plt.title("Initialization time vs accuracy")
plt.ylabel("% accuracy")
plt.xlabel("Initialization time (s)")
plt.tight_layout()
plt.subplot(234)
plt.plot(x_values, y_values_query)
plt.title("{} vs query time".format(param_name))
plt.ylabel("query time")
plt.xlabel(param_name)
plt.tight_layout()
plt.subplot(235)
plt.plot(y_values_query, y_values_accuracy)
plt.title("query time vs accuracy")
plt.ylabel("% accuracy")
plt.xlabel("query time (s)")
plt.tight_layout()
plt.show()
def evaluate_nmslib_performance(parameter, is_parameter_query, parameter_start, parameter_end, parameter_step):
nmslib_x_values = []
nmslib_y_values_init = []
nmslib_y_values_accuracy = []
nmslib_y_values_query = []
index_params = {'M': 100, 'indexThreadQty': 10, 'efConstruction': 100, 'post': 0}
query_params = {'efSearch': 100}
for x in range(parameter_start, parameter_end, parameter_step):
nmslib_x_values.append(x)
start_time = time.time()
if is_parameter_query:
query_params[parameter] = x
else:
index_params[parameter] = x
nmslib_index = NmslibIndexer(model
, index_params
, query_params)
nmslib_y_values_init.append(time.time() - start_time)
approximate_results = model.most_similar([model.wv.syn0norm[0]], topn=100, indexer=nmslib_index)
top_words = [result[0] for result in approximate_results]
nmslib_y_values_accuracy.append(len(set(top_words).intersection(exact_results)))
nmslib_y_values_query.append(avg_query_time(nmslib_index, queries=queries))
create_evaluation_graph(nmslib_x_values,
nmslib_y_values_init,
nmslib_y_values_accuracy,
nmslib_y_values_query,
parameter)
# Evaluate nmslib indexer, changing the parameter M
evaluate_nmslib_performance("M", False, 50, 401, 50)
# Evaluate nmslib indexer, changing the parameter efConstruction
evaluate_nmslib_performance("efConstruction", False, 50, 1001, 100)
# Evaluate nmslib indexer, changing the parameter efSearch
evaluate_nmslib_performance("efSearch", True, 50, 401, 100)
# Evaluate annoy indexer, changing the parameter num_tree
annoy_x_values = []
annoy_y_values_init = []
annoy_y_values_accuracy = []
annoy_y_values_query = []
for x in range(100, 401, 50):
annoy_x_values.append(x)
start_time = time.time()
annoy_index = AnnoyIndexer(model, x)
annoy_y_values_init.append(time.time() - start_time)
approximate_results = model.most_similar([model.wv.syn0norm[0]], topn=100, indexer=annoy_index)
top_words = [result[0] for result in approximate_results]
annoy_y_values_accuracy.append(len(set(top_words).intersection(exact_results)))
annoy_y_values_query.append(avg_query_time(annoy_index, queries=queries))
create_evaluation_graph(annoy_x_values,
annoy_y_values_init,
annoy_y_values_accuracy,
annoy_y_values_query,
"num_tree")
# nmslib indexer changing the parameter M, efConstruction, efSearch
nmslib_y_values_init = []
nmslib_y_values_accuracy = []
nmslib_y_values_query = []
for M in [100, 200]:
for efConstruction in [100, 200]:
for efSearch in [100, 200]:
start_time = time.time()
nmslib_index = NmslibIndexer(model,
{'M': M, 'indexThreadQty': 10, 'efConstruction': efConstruction, 'post': 0},
{'efSearch': efSearch})
nmslib_y_values_init.append(time.time() - start_time)
approximate_results = model.most_similar([model.wv.syn0norm[0]], topn=100, indexer=nmslib_index)
top_words = [result[0] for result in approximate_results]
nmslib_y_values_accuracy.append(len(set(top_words).intersection(exact_results)))
nmslib_y_values_query.append(avg_query_time(nmslib_index, queries=queries))
# Make a comparison between annoy and nmslib indexer
plt.figure(1, figsize=(12, 6))
plt.subplot(121)
plt.scatter(nmslib_y_values_init, nmslib_y_values_accuracy, label="nmslib", color='r', marker='o')
plt.scatter(annoy_y_values_init, annoy_y_values_accuracy, label="annoy", color='b', marker='x')
plt.legend()
plt.title("Initialization time vs accuracy. Upper left is better.")
plt.ylabel("% accuracy")
plt.xlabel("Initialization time (s)")
plt.subplot(122)
plt.scatter(nmslib_y_values_query, nmslib_y_values_accuracy, label="nmslib", color='r', marker='o')
plt.scatter(annoy_y_values_query, annoy_y_values_accuracy, label="annoy", color='b', marker='x')
plt.legend()
plt.title("Query time vs accuracy. Upper left is better.")
plt.ylabel("% accuracy")
plt.xlabel("Query time (s)")
plt.xlim(min(nmslib_y_values_query+annoy_y_values_query), max(nmslib_y_values_query+annoy_y_values_query))
plt.tight_layout()
plt.show() | docs/notebooks/nmslibtutorial.ipynb | RaRe-Technologies/gensim | lgpl-2.1 |
6. Work with Google word2vec files
Our model can be exported to a word2vec C format. There is a binary and a plain text word2vec format. Both can be read with a variety of other software, or imported back into gensim as a KeyedVectors object. | # To export our model as text
model.wv.save_word2vec_format('/tmp/vectors.txt', binary=False)
from smart_open import open
# View the first 3 lines of the exported file
# The first line has the total number of entries and the vector dimension count.
# The next lines have a key (a string) followed by its vector.
with open('/tmp/vectors.txt') as myfile:
for i in range(3):
print(myfile.readline().strip())
# To import a word2vec text model
wv = KeyedVectors.load_word2vec_format('/tmp/vectors.txt', binary=False)
# To export our model as binary
model.wv.save_word2vec_format('/tmp/vectors.bin', binary=True)
# To import a word2vec binary model
wv = KeyedVectors.load_word2vec_format('/tmp/vectors.bin', binary=True)
# To create and save Nmslib Index from a loaded `KeyedVectors` object
nmslib_index = NmslibIndexer(wv,
{'M': 100, 'indexThreadQty': 1, 'efConstruction': 100}, {'efSearch': 100})
nmslib_index.save('/tmp/mymodel.index')
# Load and test the saved word vectors and saved nmslib index
wv = KeyedVectors.load_word2vec_format('/tmp/vectors.bin', binary=True)
nmslib_index = NmslibIndexer.load('/tmp/mymodel.index')
nmslib_index.model = wv
vector = wv["cat"]
approximate_neighbors = wv.most_similar([vector], topn=11, indexer=nmslib_index)
# Neatly print the approximate_neighbors and their corresponding cosine similarity values
print("Approximate Neighbors")
for neighbor in approximate_neighbors:
print(neighbor)
normal_neighbors = wv.most_similar([vector], topn=11)
print("\nNormal (not Nmslib-indexed) Neighbors")
for neighbor in normal_neighbors:
print(neighbor) | docs/notebooks/nmslibtutorial.ipynb | RaRe-Technologies/gensim | lgpl-2.1 |
End of preview. Expand
in Dataset Viewer.
This is a parsed version of github-jupyter-parsed, with markdown and code pairs. We provide the preprocessing script in preprocessing.py. The data is deduplicated and consists of 451662 examples.
For similar datasets with text and Python code, there is CoNaLa benchmark from StackOverflow, with some samples curated by annotators.
- Downloads last month
- 115