uncomplicate

2017-06-08T08:39:51.635286Z

@whilo I'd need to see at least a hello-world-ish code of those ideas to be able to form an opinion.

whilo 2017-06-08T20:45:54.233185Z

what high dimensional problems can you tackle?

whilo 2017-06-08T20:46:09.238164Z

i would like to embed vision problems, but this seems fairly ambitious

2017-06-08T20:56:10.430490Z

I can see that the example takes 20.000 samples of one-dimensional distribution. That's a fairly trivial problem. See Doing Bayesian Data Analysis book - it shows many practical hierarchical models and many examples go into 50+ dimensions. All the examples compute in a fraction of second with bayadera (the same examples run for minutes with Stan, which does all fancy variational + hamiltonian mcmc etc. in C++ called from R). Anyway, for vision problems, which is perception, I do not see how anything can beat deep neural nets...

whilo 2017-06-08T21:00:01.504767Z

Right. GANs should be embeddable in generative models.

whilo 2017-06-08T21:00:45.520172Z

You asked for hello world, there are a lot more worksheets.

whilo 2017-06-08T21:00:49.521348Z

🙂

whilo 2017-06-08T21:01:53.541987Z

But you are right, that bayadera is much more focused on performance. For vision this might be very helpful.

whilo 2017-06-08T21:02:47.559570Z

I just want to point out that the work around anglican is fairly broad and innovative in regard to language embedding and composability.

whilo 2017-06-08T21:03:47.578455Z

I can totally imagine having a highly optimized bayadera model being part of it. As far as I understand the models are composable at the boundaries in general.

2017-06-08T21:04:21.588984Z

the problem is that those computations are so demanding that performance is THE metric. It doesn't matter if anglican can create more complex models if those models run in days. BTW, can you please do a quick microbenchmark and tell me how long does it take anglican to take those 20.000 samples (do not forget to unroll the sequence with doall)

2017-06-08T21:05:44.614206Z

another thing I do not like with anglican's approach is that it does not support regular clojure functions, but does some sort of model compilation that require the model code to be somehow special (i.e. not clojure)

2017-06-08T21:07:49.650480Z

when I talked about hello world I ment the demonstration of those fancy ideas (variational inference, NN, etc.) and the comparison with some baseline wrt. some metrics

whilo 2017-06-08T21:08:02.654176Z

It does allow passing in of normal clojure functions through with-primitive-procedures

whilo 2017-06-08T21:08:33.663019Z

Right, I get your performance point.

whilo 2017-06-08T21:08:58.669943Z

Btw. what do you think of edward?

2017-06-08T21:11:00.705278Z

I'm not too much into probabilistic programming, I am more interested in probabilistic data analysis, and probabilistic machine learning. these things are based on the same theory, but are not the same.

2017-06-08T21:11:27.713228Z

moreover, I look at practicality.

2017-06-08T21:11:41.717545Z

not so much after pure research of interesting things any more

whilo 2017-06-08T21:11:51.720287Z

The problem is not so much whether to do it in anglican or in bayadera, but whether to build on Clojure at all.

whilo 2017-06-08T21:11:54.721180Z

I see.

whilo 2017-06-08T21:13:05.742158Z

I agree about the programming aspect, although I think it is possible to have fast probabilistic machine learning in such a language and optimize inference with it through "compilation".

2017-06-08T21:13:20.746707Z

I mean, I am after interesting things, but I set a higher bar 🙂

whilo 2017-06-08T21:13:26.748478Z

Practically speaking machine learning has not entered the programmers toolbox yet.

2017-06-08T21:13:28.749164Z

it also have to solve real problemt

2017-06-08T21:13:32.750190Z

problems.

2017-06-08T21:13:45.753773Z

most of those research project demonstrate toy problems

whilo 2017-06-08T21:13:48.754684Z

From that direction anglican might be much more approachable to embed in some small problems, than going full machine learning.

whilo 2017-06-08T21:13:59.757912Z

I agree.

whilo 2017-06-08T21:14:17.762597Z

But I am in a group who does heavy vision problems and there it is the opposite.

whilo 2017-06-08T21:14:30.766164Z

Probabilistic programming doesn't cut it for these problems.

2017-06-08T21:14:34.767388Z

yep, and I want to create such toolbox

whilo 2017-06-08T21:14:53.772616Z

Or more precisely a bayesian approach.

2017-06-08T21:14:54.772982Z

of course, because vision is about perception

2017-06-08T21:15:01.774799Z

and not about logic

whilo 2017-06-08T21:15:14.778530Z

I meant the cost to compute uncertainties.

2017-06-08T21:15:46.787801Z

probabilistic techniques might be interesting next layer, that could do some reasoning on the output of the vision layer

whilo 2017-06-08T21:16:03.792749Z

Yes, that is what most people do nowadays.

whilo 2017-06-08T21:16:13.795628Z

They use some CNN and just use it as a feature extractor.

2017-06-08T21:16:39.802934Z

yep.

whilo 2017-06-08T21:16:51.806355Z

(and it works very very well) 🙂

whilo 2017-06-08T21:17:52.823764Z

btw. next on my reading list: https://arxiv.org/pdf/1703.04977.pdf

whilo 2017-06-08T21:19:08.845084Z

For the GMM the 20000 samples took 100secs on my laptop (unsurprisingly).

whilo 2017-06-08T21:19:40.853833Z

MC methods are inefficient in general and not respected much in machine learing (at least from my environment)

whilo 2017-06-08T21:20:27.866847Z

This is their newest take: https://arxiv.org/abs/1705.10306

2017-06-08T21:20:29.867551Z

And, that is the Gaussian distribution, which is the easiest distribution to sample from (after the Uniform)

whilo 2017-06-08T21:21:47.889461Z

I know. But I don't think bayadera couldn't be described by the anglican syntax and framework. I think the big success of NNs besides initial breakthroughs is mostly due to the fact that modern toolkits allow easy composition and modelling.

2017-06-08T21:21:48.889815Z

Now, Bayadera can take much, much, much, more (I forgot how many) samples from Gaussian in a few dozens milliseconds, and most of this time is the communication with the GPU.

2017-06-08T21:22:20.898537Z

But consider this: 100 secs for the simplest hello world that you could find

2017-06-08T21:22:25.900097Z

how useful is that?

2017-06-08T21:22:34.902584Z

no matter what features are there?

whilo 2017-06-08T21:22:54.907953Z

It was the GMM on iris, it was not the worksheet i have sent you. But it is still sample.

whilo 2017-06-08T21:23:30.918291Z

simple

2017-06-08T21:23:40.921042Z

Bayadera gives you ordinary clojure functions. Why wouldn't you be able to compose them?

whilo 2017-06-08T21:27:04.977153Z

I see. The problem is that I can barely convince anybody to use Clojure. For Bayadera to be attractive it would help if it would be part of a bigger community. Clojure in machine learning is still a very hard sell. I can probably use something on my own, but it will be difficult to attract colleagues. Anglican is not much better in that regard, but the few really nice projects that are out there feel very isolated and fragmented. My colleague would like to go with edward, I guess, since it is sponsored by OpenAI and built on top of tensorflow (although he doesn't like tensorflow in particular).

whilo 2017-06-08T21:28:00.991865Z

What I like about Anglican is that I can see people using it for small data problems and inference. If this is possible with Bayadera as well, I am totally fine with it.

whilo 2017-06-08T21:28:49.005070Z

With people I mean everyday Clojure developers without a GPU and a background in data science.

whilo 2017-06-08T21:29:04.009375Z

This would allow to grow a community.

2017-06-08T21:29:51.021898Z

That's why I don't like to bump people to use my (or other) tools. I'm OK with competition using inferior tools 🙂 OTOH, the best way to convince people to use some thechnology is to build useful stuff with it. When they see the result of your work, they'll ask you to tell them what you did (provided that you did something awesome, of course).

2017-06-08T21:30:56.040489Z

Now, it is difficult to convince people to use Clojure for ML, when there are lots of pies-in-the-sky talk, but Clojure tools like Incanter are a joke.

2017-06-08T21:31:52.056108Z

I think that the GPU is essential for ML

2017-06-08T21:32:12.061724Z

For most methods, at least

2017-06-08T21:32:40.069514Z

And the theory has to be learned to some degree

whilo 2017-06-08T21:33:09.077051Z

Hmm, yes. You are right about proving with results and the GPU. Anglican can leverage the GPU by embedding LSTMs or VAEs for its proposal distribution, just to point out.

2017-06-08T21:33:37.084939Z

Whoever hopes that they will be able to do ML with the level of knowledge required for Web apps and no maths, will spent years circling around punching other people's tutorials

2017-06-08T21:34:06.092836Z

I'd like to see some benchmarks

2017-06-08T21:37:23.144414Z

BTW Bayadera IS useful for small data problems, and I doubt it is useful for big data problems. That goes to Bayesian methods in general.

2017-06-08T21:37:39.148638Z

But, small data usually means big computation

2017-06-08T21:37:48.150985Z

and Bayadera is all about that 🙂

whilo 2017-06-08T21:38:54.168141Z

https://arxiv.org/pdf/1705.10306.pdf Section 5.4 has a 4096 dimensional sampling space.

whilo 2017-06-08T21:39:19.174258Z

But no statement over training time or inference.

whilo 2017-06-08T21:39:24.175559Z

Just model quality.

2017-06-08T21:41:00.199910Z

that's the problem with most papers. they count number of steps, without regard how much one step costs wrt computation, memory access, and parallelization

whilo 2017-06-08T21:41:25.206385Z

I agree that you need maths. But doing maths before seeing what you can do with machine learning can turn many people off. Esp. when you do probabilistic modelling, you need to do a lot of math, much more than for NNs.

whilo 2017-06-08T21:41:40.210162Z

Yes, right.

whilo 2017-06-08T21:42:16.219485Z

But I think combining MC samplers with NNs that way might be a very good idea.

whilo 2017-06-08T21:42:47.227439Z

Inference will become a lot cheaper once the NN is trained (they coin it "compiled").

2017-06-08T21:43:08.232658Z

I don't think so - NNs also require maths to understand what you do, it's only that there are popular frameworks with ready made recipes, that work well for handful of problems (mostly vision and nlp). however, what if you do not work with vision and nlp?

whilo 2017-06-08T21:43:16.234615Z

What do you plan to show off with Bayadera? 😛

whilo 2017-06-08T21:43:48.242431Z

Yes, I like Bayesian approaches and generative models.

whilo 2017-06-08T21:43:55.244029Z

I think they generalize a lot better.

whilo 2017-06-08T21:44:08.247296Z

NNs just need you to understand gradients.

whilo 2017-06-08T21:44:20.250192Z

No statistics.

2017-06-08T21:44:42.255493Z

Well, I already have (for more than a year) large speedup over Stan, but I am not even eager to publish that. I plan to work on some commercial applications, so I do not even want to show off the technology itself, but the product.

whilo 2017-06-08T21:44:59.259757Z

And you are right, that is also my problem. As long as the toolboxes for black-box bayesian inference are complicated, they will never be as popular.

whilo 2017-06-08T21:45:23.265819Z

I understand. Might be a good plan.

2017-06-08T21:46:15.278536Z

The biggest "doh" is that bayesian stuff isn't even that complicated, especially when you use the tools that someone other made.

whilo 2017-06-08T21:46:25.280975Z

Why have you presented at bobkonf then?

2017-06-08T21:46:29.282057Z

Stan is used by social scientists and biologists

whilo 2017-06-08T21:47:18.293915Z

Yes, that is why I like the blackbox approaches where you can specify a generative process and then have reasonable inference per default.

whilo 2017-06-08T21:47:49.301709Z

I think machine learning should stop being a niche thing for highly skilled specialists.

2017-06-08T21:48:28.311254Z

The problem with programmers is that sometimes they... well, almost like physicists. But, I do not even care. I open-sourced the technology because at least some people can "get it" (and some really do get it and use it) and I might get valuable feedback and even some contributions. However, I do not want to beg anyone to use it and why would I?

whilo 2017-06-08T21:48:29.311278Z

I have been motivated by LDA 4 years ago, it was a very intuitive generative model and easy to understand even without heavy math knowledge back then.

2017-06-08T21:48:56.318122Z

Did it solve practical problems?

whilo 2017-06-08T21:49:08.320962Z

You are right. I am thankful that you take the time to argue with me.

whilo 2017-06-08T21:49:27.325824Z

LDA?

whilo 2017-06-08T21:49:30.326744Z

Definitely

whilo 2017-06-08T21:50:06.336115Z

You probably know it, it is known as topic modelling to the industry.

2017-06-08T21:50:35.343290Z

I don't use that. It is used in NLP?

whilo 2017-06-08T21:50:41.344763Z

Yes

2017-06-08T21:50:51.347409Z

I don't do NLP, that's the thing 🙂

whilo 2017-06-08T21:51:13.353079Z

You can get the topics generating a corpus of texts and the distribution of topics per document in an unsupervised fashion from it.

whilo 2017-06-08T21:51:18.354480Z

It is really handy.

whilo 2017-06-08T21:51:34.358362Z

(even if just two browse your local paper collection 🙂 )

2017-06-08T21:51:55.363667Z

Did you have any application/business ideas around that or more like research curiosity?

whilo 2017-06-08T21:52:07.366780Z

It was at university.

whilo 2017-06-08T21:52:39.374484Z

I definitely has business value and is heavily marketed already.

whilo 2017-06-08T21:52:48.376751Z

The original paper is from Blei et al. 2003

whilo 2017-06-08T21:53:09.382208Z

Something like topic modelling for vision would be nice.

2017-06-08T21:53:41.390252Z

Are you still looking for the thesis topic, or you are set with something?

whilo 2017-06-08T21:54:18.399354Z

I am trying to be set. It is frustrating. I would like to at least work probablistically and not just throw NNs at some thing.

whilo 2017-06-08T21:54:28.401858Z

So, no.

whilo 2017-06-08T21:54:34.403389Z

I was with my supervisor today.

whilo 2017-06-08T21:54:50.407268Z

They are very focused on vision.

whilo 2017-06-08T21:55:57.423851Z

I don't want to just do some topic they throw at me. I like to be motivated by myself, but this is not working well this time.

2017-06-08T21:56:26.430994Z

Did you see the (old-ish) book by Bishop called Pattern recognition and machine learning?

2017-06-08T21:56:37.433648Z

It's from the pre-DL era (2006)

whilo 2017-06-08T21:56:39.433995Z

Yes, I worked through parts of it.

whilo 2017-06-08T21:56:56.438303Z

Mostly first 100 pages, EM and variational inference.

2017-06-08T21:57:08.441181Z

but the book is probabilistically-oriented, and he discusses the probabilistic perspective of NNs

2017-06-08T21:57:18.443810Z

and similarity to bayesian nets

2017-06-08T21:57:59.454030Z

but the problem is that DL is so successfull with vision and perception in general that it is a slim chance that you'll get something with bayesian methods

whilo 2017-06-08T21:58:08.455836Z

Yes.

whilo 2017-06-08T21:58:16.457840Z

It is also not clear what the uncertainty buys you.

2017-06-08T21:58:16.458004Z

they are simply the hammer for another kind of nails

whilo 2017-06-08T21:58:22.459129Z

It definitely costs computation.

2017-06-08T21:58:40.463421Z

it does not buy you anything if you have enough data

2017-06-08T21:58:51.466127Z

but in vision you usually have lots of data 🙂

whilo 2017-06-08T21:59:02.468997Z

Well, the paper I have first posted is from a guy (Yarin Gal), who showed that droput regularization turns the neural net into a bayesian neural net (or a GP).

2017-06-08T21:59:18.472589Z

maybe bayesian techniques could be useful when you have extremely scarce visual information

2017-06-08T21:59:24.474015Z

but is there such domain?

whilo 2017-06-08T21:59:26.474367Z

So you can use the uncertainty on the output by running the inference several times in a meaningful way.

2017-06-08T21:59:54.481134Z

what is your prior there

2017-06-08T21:59:56.481358Z

?

whilo 2017-06-08T22:00:15.486478Z

An example from Anglican was to have generative model for captchas and theirs was state of the art, cracking all of them.

whilo 2017-06-08T22:01:16.501526Z

Your prior is a GP prior. That corresponds to a normal distribution over the weight matrices.

2017-06-08T22:02:54.524418Z

wait, wait. I meant, you, as a human, set some prior that describe your current (pre-data) knowledge. How do you decide on that?

2017-06-08T22:06:24.571448Z

That article looks interesting

2017-06-08T22:06:33.573352Z

I put it in my bookmarks

whilo 2017-06-08T22:06:47.576217Z

It depends on whether you parametrize the normal distribution again.

2017-06-08T22:06:59.578720Z

Hope I'll have time to go through it more attentively in several months 🙂

whilo 2017-06-08T22:07:05.580115Z

Hehe

2017-06-08T22:07:31.585804Z

Ah, but why does it have to be the Normal? 🙂

2017-06-08T22:07:54.590419Z

There are many kinds of random processes, and many kinds of distributions 🙂

2017-06-08T22:08:15.594821Z

I understand that in vision, Gaussian might be the thing.

2017-06-08T22:08:54.602783Z

But generally, let's say I am trying to estimate when people call the call center, or something like that

2017-06-08T22:09:27.609876Z

Or the risk of giving out loans

whilo 2017-06-08T22:09:41.612719Z

You pick a kernel function to calculate the covariance matrix of the gaussian K

whilo 2017-06-08T22:09:50.614756Z

and then the prior is GP(·|0, K)

2017-06-08T22:09:55.615822Z

Or some general risk -> Bayesian methods are really good match for measuring risk

whilo 2017-06-08T22:10:13.619795Z

Hmm, no that makes no sense

2017-06-08T22:10:23.621927Z

?

whilo 2017-06-08T22:15:43.687630Z

The prior knowledge flows into the kernel function.

whilo 2017-06-08T22:15:49.688707Z

for the covariance matrix

whilo 2017-06-08T22:16:30.697032Z

from that you generate a function that has some prior set on the uncertainty of the measurement.

whilo 2017-06-08T22:16:58.702631Z

All distributions are Gaussian, hence the whole thing is Gaussian again.

whilo 2017-06-08T22:18:28.721177Z

But the kernel function can do arbitrariy complicated things, it can be a NN.

2017-06-08T22:18:34.722247Z

Note that the model is pretty sure about the parameters, and they are not very probabilistic 🙂

2017-06-08T22:18:36.722707Z

anyway

2017-06-08T22:18:51.725548Z

I get that, but, given an unknown problem,

whilo 2017-06-08T22:19:06.728767Z

I know. I don't like kernel methods very much.

2017-06-08T22:19:16.730542Z

how do you decide it is a good fit to be described by Gaussian likelihood,

2017-06-08T22:19:36.734901Z

and how do you transfer your prior knowledge to the parameters?

2017-06-08T22:19:54.738540Z

Yep, many "bayesian" methods are not that much bayesian

2017-06-08T22:20:03.740322Z

which doesn't mean they are not useful

whilo 2017-06-08T22:20:30.745816Z

There are deep GPs now btw. Where the joint probability is not Gaussian any more.

whilo 2017-06-08T22:20:38.747407Z

That is what the blog post talks about as well.

2017-06-08T22:20:57.751226Z

Anyway, this doesn't help you in choosing the topic 🙂

2017-06-08T22:21:17.755281Z

What outcome is expected of you?

2017-06-08T22:21:37.758964Z

A number of published papers, or something else?

2017-06-08T22:22:22.767467Z

Does it have to be a part of a narrow EU/DE/industry-backed project or you are more free in choosing the area as long as it is vision?

whilo 2017-06-08T22:22:49.772748Z

Solving a good vision problem, something practical. Not just doing maths or playing around.

whilo 2017-06-08T22:23:01.774958Z

Although I want to improve my math skill still.

whilo 2017-06-08T22:23:18.778106Z

I think targeting a good paper as a result would be reasonable.

whilo 2017-06-08T22:23:23.779203Z

But not required.

2017-06-08T22:23:30.780633Z

Solving to be the best in the world or just good enough?

2017-06-08T22:23:55.785290Z

Really? The paper is not required? You have it easy 🙂

whilo 2017-06-08T22:24:04.786933Z

I think good enough would be ok.

whilo 2017-06-08T22:24:10.788063Z

I am not sure. I want the paper anyway.

2017-06-08T22:24:16.789152Z

Good enough is ok? Even better 🙂

2017-06-08T22:24:25.790921Z

Why worry then?

whilo 2017-06-08T22:25:04.798305Z

It would be good if we can show that bayesian modelling can work for vision problems, atm. weakly supervised problems are interesting.

whilo 2017-06-08T22:25:12.799861Z

(to the group)

2017-06-08T22:25:31.803724Z

Do you have some problem where data is really, really poor and DL does not work well?

whilo 2017-06-08T22:25:56.808565Z

I would like to do something good and know that I can work well with the methods.

2017-06-08T22:26:16.812401Z

That's a Herculean task...

whilo 2017-06-08T22:26:39.816998Z

I am also thinking about doing a PhD.

2017-06-08T22:26:52.819583Z

DL (and most of ML) fields seems to me as many soothsayers throwing bones around and reading star constellations

whilo 2017-06-08T22:27:06.822308Z

Hehe

2017-06-08T22:27:10.823015Z

You are not already on that path?

2017-06-08T22:27:20.825057Z

Are you in MSc?

whilo 2017-06-08T22:27:25.825969Z

Yes.

whilo 2017-06-08T22:27:35.828011Z

This is my master thesis.

2017-06-08T22:27:58.832380Z

Ah, in that case, why are you talking such a heavy task?

2017-06-08T22:28:32.839565Z

Isn't it more appropriate for that level to take something that has been researched well and make a good implementation?

whilo 2017-06-08T22:28:47.842386Z

Good question. Maybe I have stayed too long in the math department and have complexes now.

2017-06-08T22:29:03.845556Z

That's why I asked you about the paper.

whilo 2017-06-08T22:29:04.845846Z

I have done this in my bachelor thesis already.

2017-06-08T22:29:19.848668Z

What you are talking here is too much for master's thesis.

2017-06-08T22:29:25.849718Z

Especially wrt time.

whilo 2017-06-08T22:29:54.855106Z

Right.

whilo 2017-06-08T22:30:25.861607Z

Well, I mostly see paper's and thesis from state of the art people in the field and they are really good.

2017-06-08T22:30:47.866081Z

But these people are doing post-doc research full time.

2017-06-08T22:31:07.869932Z

And often have worked on those problems for years or even decades.

2017-06-08T22:31:23.873004Z

DL people were doing that thing since 89s

2017-06-08T22:31:26.873520Z

80s

2017-06-08T22:31:33.874619Z

when it was not cool

2017-06-08T22:31:37.875390Z

and not very useful 🙂

whilo 2017-06-08T22:31:43.876627Z

Right.

2017-06-08T22:32:07.881216Z

So you can try to do the same for master's if you have 10 years for that 🙂

whilo 2017-06-08T22:32:25.884531Z

I don't. I need to break the problem down and do something focused.

2017-06-08T22:32:32.885950Z

but it is better to do some cool exploratory implementation as an introduction to phd

2017-06-08T22:32:46.888457Z

or that is what I'd advise my students 🙂

2017-06-08T22:33:39.898521Z

although they are all working programmers, since in Serbia there is little opportunity to fund students to work on research so they are all working full-time in industry cranking code, and do their studies part-time

whilo 2017-06-08T22:33:42.899031Z

That is a reasonable advice.

whilo 2017-06-08T22:34:05.903376Z

That is tough.

2017-06-08T22:34:13.904720Z

yep

whilo 2017-06-08T22:34:24.906797Z

I have worked part-time with Clojure over the last half year, but I was lucky to do so with friends.

whilo 2017-06-08T22:34:33.908366Z

It payed not very well, but it was ok.

2017-06-08T22:34:51.911690Z

cool. what are you making?

whilo 2017-06-08T22:35:08.914862Z

it was an app prototype for a client who wants something like yelp.

whilo 2017-06-08T22:35:18.916883Z

although i think he doesn't know exactly what he wants.

whilo 2017-06-08T22:35:49.922626Z

but we could do datomic with fulltext-search and google places integration in the backend and two single page apps and an android and ios app.

2017-06-08T22:35:58.924328Z

well, at least you polished your Clojure and were paid for that 🙂

whilo 2017-06-08T22:36:00.924713Z

it is not finished yet, but clojure was reasonable.

whilo 2017-06-08T22:36:06.925863Z

yes.

whilo 2017-06-08T22:36:19.928292Z

we also stayed in our deadlines, which was good to see.

whilo 2017-06-08T22:36:50.934275Z

i need to take the dog for a walk

2017-06-08T22:37:01.936249Z

enjoy 🙂

whilo 2017-06-08T22:37:12.938377Z

really nice talking to you, hopefully we can continue this soon

whilo 2017-06-08T22:37:20.939749Z

can bayadera run without a GPU atm.=?

whilo 2017-06-08T22:37:29.941446Z

i just have this laptop right now

2017-06-08T22:37:32.941965Z

no.

whilo 2017-06-08T22:37:36.942804Z

ok

whilo 2017-06-08T22:37:46.944480Z

no problem

whilo 2017-06-08T22:37:49.945192Z

i can get a gpu

2017-06-08T22:38:25.951652Z

atm it requires AMD GPU that supports OpenCL 2.0, but I'm also probably giving it CUDA backend soon-ish

2017-06-08T22:38:36.953635Z

and MAYBE a CPU backend

2017-06-08T22:38:52.956500Z

But that depends on inspiration 🙂

whilo 2017-06-08T22:39:00.958120Z

no pressure. has nvidia reasonable opencl support?

2017-06-08T22:39:08.959403Z

no. 1.2

whilo 2017-06-08T22:39:14.960368Z

pfff

whilo 2017-06-08T22:39:33.964027Z

i follow linus' comments... 😉

2017-06-08T22:39:50.967061Z

although my card, R9 290X, which is quite a beast, probably costs something like 100 EUR in Germany now (second-hand)

2017-06-08T22:40:17.972054Z

Because it's couple of generations old, but was top of the line 3 years ago.

2017-06-08T22:40:29.974390Z

So it might be a modest investments

2017-06-08T22:40:56.979294Z

While my Nvidia 1080 cost around 1000 EUR here less than a year ago

2017-06-08T22:41:07.981383Z

and is only 30% faster

whilo 2017-06-08T22:41:38.987073Z

Hmm, interesting.

whilo 2017-06-08T22:41:48.988766Z

Yes, nvidia was always fairly expensive.

whilo 2017-06-08T22:42:03.991440Z

At least when I last had a look at perf 5 years ago or so.

2017-06-08T22:42:39.998566Z

Nvidia's main strength is its suite of hand-tuned libraries, cuBLAS, cuDNN etc.

whilo 2017-06-08T22:42:52.001228Z

Have you integrated NNs (e.g. pretrained) in your pipeline yet?

whilo 2017-06-08T22:43:03.003197Z

Yes, I think so too.

2017-06-08T22:43:05.003493Z

no. I'm not that into nns

2017-06-08T22:43:27.007690Z

Everybody's doing that, and they are pretty optimized the field

whilo 2017-06-08T22:43:35.009331Z

I see. There are tons of pretrained models out there though, which could be helpful for industry applications.

whilo 2017-06-08T22:43:39.010198Z

E.g. pretrained CNNs.

2017-06-08T22:43:44.011035Z

I want to work in an uncluttered niche

whilo 2017-06-08T22:43:57.013061Z

I understand that very well.

whilo 2017-06-08T22:44:50.023021Z

Well, this should be mostly plumbing, nothing critical. pretrained models can be applied to data separately anyway.

whilo 2017-06-08T22:44:55.024038Z

What do you use to store tensors?

whilo 2017-06-08T22:45:03.025688Z

I used hdf5 in my bachelor thesis.

whilo 2017-06-08T22:45:09.027058Z

Clojure support is a bit weak.

whilo 2017-06-08T22:45:21.029240Z

I had to tweak the library for multiple dims.

2017-06-08T22:45:23.029716Z

Nothing now. I don't need them for bayesian stuff (yet).

2017-06-08T22:45:45.033833Z

When I need them, I'll add them to neanderthal...

2017-06-08T22:46:17.039597Z

With a Intel-based native backend, and cuDNN GPU backend probably

whilo 2017-06-08T22:46:46.044840Z

Ok

2017-06-08T22:46:47.045147Z

as for storing, I leave that to the users for now. They can implement their own transfer methods to/from storage

whilo 2017-06-08T22:47:02.047758Z

Yes, I wouldn't pack that into the libraries.

2017-06-08T22:47:07.048503Z

your dog's getting impatient 🙂

whilo 2017-06-08T22:47:15.050185Z

It really sucks that all people in python use numpy pickeling.

whilo 2017-06-08T22:47:28.052472Z

That is right 🙂

whilo 2017-06-08T22:47:30.052968Z

Cu soon

2017-06-08T22:47:35.053862Z

bye