If performance is a major factor you should avoid using cells, loops or cellfun/arrayfun. It's usually much quicker to use a vector operation (assuming this is possible).
The code below expands on Werner's add example with standard array loop and array operations.
The results are:
- Cell Loop Time - 0.1679
- Cellfun Time - 2.9973
- Loop Array Time - 0.0465
- Array Time - 0.0019
Code:
nTimes = 1000;
nValues = 1000;
myCell = repmat({0},1,nValues);
output = zeros(1,nValues);
% Basic operation
tic;
for k=1:nTimes
for m=1:nValues
output(m) = myCell{m} + 1;
end
end
cell_loop_timeAdd=toc;
fprintf(1,'Cell Loop Time %0.4f\n', cell_loop_timeAdd);
tic;
for k=1:nTimes
output = cellfun(@(in) in+1,myCell);
end
cellfun_timeAdd=toc;
fprintf(1,'Cellfun Time %0.4f\n', cellfun_timeAdd);
myData = repmat(0,1,nValues);
tic;
for k=1:nTimes
for m=1:nValues
output(m) = myData(m) + 1;
end
end
loop_timeAdd=toc;
fprintf(1,'Loop Array Time %0.4f\n', loop_timeAdd);
tic;
for k=1:nTimes
output = myData + 1;
end
array_timeAdd=toc;
fprintf(1,'Array Time %0.4f\n', array_timeAdd);
Answer from grantnz on Stack OverflowIf performance is a major factor you should avoid using cells, loops or cellfun/arrayfun. It's usually much quicker to use a vector operation (assuming this is possible).
The code below expands on Werner's add example with standard array loop and array operations.
The results are:
- Cell Loop Time - 0.1679
- Cellfun Time - 2.9973
- Loop Array Time - 0.0465
- Array Time - 0.0019
Code:
nTimes = 1000;
nValues = 1000;
myCell = repmat({0},1,nValues);
output = zeros(1,nValues);
% Basic operation
tic;
for k=1:nTimes
for m=1:nValues
output(m) = myCell{m} + 1;
end
end
cell_loop_timeAdd=toc;
fprintf(1,'Cell Loop Time %0.4f\n', cell_loop_timeAdd);
tic;
for k=1:nTimes
output = cellfun(@(in) in+1,myCell);
end
cellfun_timeAdd=toc;
fprintf(1,'Cellfun Time %0.4f\n', cellfun_timeAdd);
myData = repmat(0,1,nValues);
tic;
for k=1:nTimes
for m=1:nValues
output(m) = myData(m) + 1;
end
end
loop_timeAdd=toc;
fprintf(1,'Loop Array Time %0.4f\n', loop_timeAdd);
tic;
for k=1:nTimes
output = myData + 1;
end
array_timeAdd=toc;
fprintf(1,'Array Time %0.4f\n', array_timeAdd);
I will add one answer with the results that I tested myself, but I would be glad if people contribute with their knowledge, this is just a simple test I've made.
I've tested the following conditions with cell size of 1000 and 1000 loops (results on total time, and I would probably have to run more than 1000 times, because I am having a little fluctuation on the results, but anyway, this is not a scientific article):
- Basic operation (sum)
- Simple for loop: 0.2663 s
- cellfun: 9.4612 s
- String Operation (strcmp)
- Simple for loop: 1.3124 s
- cellfun: 11.8099 s
- Built-in (isempty)
- Simple for loop: 8.9042 s
- cellfun (string input -> see this reference): 0.0105 s
- cellfun (fcn handle input -> see this reference): 0.9209 s
- Non-uniform (regexp)
- Simple for loop: 24.2157 s
- cellfun (string input): 44.0424 s
So, it seems that cellfun with anonymous function calls are slower than a simple for loop, but if you will use a builtin matlab method, do it with cellfun and use it with the string quotation. This is not necessarily true for all cases, but at least for the tested functions.
The implemented test code (I am far from being an optimization specialist, so here is the code in case I did something wrong):
function ...
[loop_timeAdd,cellfun_timeAdd,...
loop_timeStr,cellfun_timeStr,...
loop_timeBuiltIn,cellfun_timeBuiltInStrInput,...
cellfun_timeBuiltyInFcnHandle,...
loop_timeNonUniform,cellfun_timeNonUniform] ...
= test_cellfun(nTimes,nCells)
myCell = repmat({0},1,nCells);
output = zeros(1,nCells);
% Basic operation
tic;
for k=1:nTimes
for m=1:nCells
output(m) = myCell{m} + 1;
end
end
loop_timeAdd=toc;
tic;
for k=1:nTimes
output = cellfun(@(in) in+1,myCell);
end
cellfun_timeAdd=toc;
% String operation
myCell = repmat({'matchStr'},1,nCells); % Add str that matches
myCell(1:2:end) = {'dontMatchStr'}; % Add another str that doesnt match
output = zeros(1,nCells);
tic;
for k=1:nTimes
for m=1:nCells
output(m) = strcmp(myCell{m},'matchStr');
end
end
loop_timeStr=toc;
tic;
for k=1:nTimes
output = cellfun(@(in) strcmp(in,'matchStr'),myCell);
end
cellfun_timeStr=toc;
% Builtin function (isempty)
myCell = cell(1,nCells); % Empty
myCell(1:2:end) = {0}; % not empty
output = zeros(1,nCells);
tic;
for k=1:nTimes
for m=1:nCells
output(m) = isempty(myCell{m});
end
end
loop_timeBuiltIn=toc;
tic;
for k=1:nTimes
output = cellfun(@isempty,myCell);
end
cellfun_timeBuiltyInFcnHandle=toc;
tic;
for k=1:nTimes
output = cellfun('isempty',myCell);
end
cellfun_timeBuiltInStrInput=toc;
% Builtin function (isempty)
myCell = repmat({'John'},1,nCells);
myCell(1:2:end) = {'Doe'};
output = cell(1,nCells);
tic;
for k=1:nTimes
for m=1:nCells
output{m} = regexp(myCell{m},'John','match');
end
end
loop_timeNonUniform=toc;
tic;
for k=1:nTimes
output = cellfun(@(in) regexp(in,'John','match'),myCell,...
'UniformOutput',false);
end
cellfun_timeNonUniform=toc;
Hi everyone,
Lately I am using functions like cellfun, arrayfun, etc. all the time to avoid writing loops. I was wondering if this is a good practise.
Is it better or simple loop which is much easier to write and read is a better approach?
In addition a for loop can run in parallel later.
I have been writing some matlab code and I've found myself using arrayfun to elegantly perform certain operations without a for loop. (similar to list comprehension in Python).
However, I started to think about the speed of it. so I made a contrived example of squaring a large matrix. (my actual uses are a bit more fancy). See the code below if you're interested
The final times were:
| method | time (s) |
|---|---|
| array_fun | 5.1231 |
| array_fun_fixed | 5.0522 |
| pre_loop | 0.0343 |
| non_pre_loop | 0.3276 |
| pre_loop_fun | 0.5520 |
| pre_loop_fun_each | 71.9136 |
| non_pre_loop_fun | 0.8565 |
| direct | 9.1100e-04 |
I expected the direct would be the fastest. And I expected the preallocated loops to be faster than the dynamically sized ones. I also expected the one one where it creates the function each time to be the slowest.
And, I kind of expected for an anonymous function to be slower than direct manipulation. But what really surprised me was the (a) the HUGE slowdown of arrayfun and (b) that it was (much) slower than a loop calling the same function. It seems as though fixing the function does not matter, but arrayfun still underperforms a loop.
If anything, I would expect arrayfun to be on par with a pre-allocated loop. Assuming UniformOutput is set to false (default), Matlab knows the final size of the returned array. Furthermore, it should just be looping over it in a pretty clear manner.
Any thoughts or insights into the overhead and methods of this?
Thanks
Code
The script I ran:
clear
R = rand(1000);
% Arrafun with anonymous function
clear BLA
tic
BLA = arrayfun(@(n) n^2,R);
T.array_fun = toc;
clear BLA
tic
FUN = @(n) n^2;
BLA = arrayfun(FUN,R);
T.array_fun_fixed = toc;
% Preallocated loop
clear BLA
tic
BLA = zeros(size(R));
for ii = 1:size(R,1)
for jj = 1:size(R,2)
BLA(ii,jj) = R(ii,jj)^2;
end
end
T.pre_loop = toc;
% Non-Preallocated loop
clear BLA
tic
for ii = 1:size(R,1)
for jj = 1:size(R,2)
BLA(ii,jj) = R(ii,jj)^2;
end
end
T.non_pre_loop = toc;
% Preallocated loop with with anonymous function
clear BLA
tic
fun = @(n) n^2;
BLA = zeros(size(R));
for ii = 1:size(R,1)
for jj = 1:size(R,2)
BLA(ii,jj) = fun(R(ii,jj));
end
end
T.pre_loop_fun = toc;
% Preallocated loop with with anonymous function EACH TIME
clear BLA
tic
BLA = zeros(size(R));
for ii = 1:size(R,1)
for jj = 1:size(R,2)
fun = @(n) n^2;
BLA(ii,jj) = fun(R(ii,jj));
end
end
T.pre_loop_fun_each = toc;
% Non-Preallocated loop with with anonymous function
clear BLA
tic
fun = @(n) n^2;
for ii = 1:size(R,1)
for jj = 1:size(R,2)
BLA(ii,jj) = fun(R(ii,jj));
end
end
T.non_pre_loop_fun = toc;
% Direct
clear BLA
tic
BLA = R.^2;
T.direct = toc;Looking around on StackOverflow, it seems like this is recognized behavior. But that doesn't mean there aren't cases where arrayFun can perform better.
On my machine, my times are roughly twice yours for your code. But consider this example instead:
x = gpuArray(rand(100,100,3));
tic
z = arrayfun(@(x) x^2, x);
toc
z = gpuArray(zeros(length(x),1));
tic
for i=1:length(x(:))
z(i) = x(i).^2;
end
toc
Elapsed time is 0.011370 seconds.
Elapsed time is 17.763806 seconds.
Arrayfun's assumptions that allow for parallel code leads to a huge jump in performance in that use case (mostly because the for loop approach isn't great for running on the GPU). Not sure if it's even a fair comparison, frankly.
On Matlab Central, one poster suggests avoiding arrayfun if you're not planning on using a GPU: https://www.mathworks.com/matlabcentral/answers/144344-in-my-code-arrayfun-slower-than-for-loop
With the exception of being on a GPU, arrayfun will most likely often be slower than a for-loop and harder to read. It's just a less flexible more complex for-loop.
Personally, I'd recommend against it at all cost unless you're targeting a GPU.
Obviously, my for loop on a GPU is much slower, and the GPU arrayfun call is roughly the same speed as your "direct" method on my machine. For more complex examples, I'd expect the GPU approach to really show its stuff.
i enjoy these types of experiments. some comments on your benchmarking methodology.
-
i would either use matlabs built-in 'timeit' function or take an average of many iterations of the same thing. several results appear in the noise of a standard deviation.
-
since you're benchmarking, you want to ensure matlabs not getting in the way by trying to be helpful. it does all sorts of voodoo behind the scene to anticipate and optimize execution. that is, use 'clear all' between each test.
-
goes without saying that you're benchmarking is at the mercy of the os, so the more you can do to reduce interrupts the better.
last, i don't think 'arrayfun' is intended for the usage you demonstrated. matrix operations make no sense for arrayfun. but imagine you had an array of 1s and 0s and you wanted to test whether a given 1 was surrounded by 0s. that's the kind of thing 'arrayfun' would make easier, albeit not necessarily faster.
Hi I'm currently doing a assignment and I'm just wonder which one would be quicker. Cellfun or for loop
You can get the idea by running other versions of your code. Consider explicitly writing out the computations, instead of using a function in your loop
tic
Soln3 = ones(T, N);
for t = 1:T
for n = 1:N
Soln3(t, n) = 3*x(t, n)^2 + 2*x(t, n) - 1;
end
end
toc
Time to compute on my computer:
Soln1 1.158446 seconds.
Soln2 10.392475 seconds.
Soln3 0.239023 seconds.
Oli 0.010672 seconds.
Now, while the fully 'vectorized' solution is clearly the fastest, you can see that defining a function to be called for every x entry is a huge overhead. Just explicitly writing out the computation got us factor 5 speedup. I guess this shows that MATLABs JIT compiler does not support inline functions. According to the answer by gnovice there, it is actually better to write a normal function rather than an anonymous one. Try it.
Next step - remove (vectorize) the inner loop:
tic
Soln4 = ones(T, N);
for t = 1:T
Soln4(t, :) = 3*x(t, :).^2 + 2*x(t, :) - 1;
end
toc
Soln4 0.053926 seconds.
Another factor 5 speedup: there is something in those statements saying you should avoid loops in MATLAB... Or is there really? Have a look at this then
tic
Soln5 = ones(T, N);
for n = 1:N
Soln5(:, n) = 3*x(:, n).^2 + 2*x(:, n) - 1;
end
toc
Soln5 0.013875 seconds.
Much closer to the 'fully' vectorized version. Matlab stores matrices column-wise. You should always (when possible) structure your computations to be vectorized 'column-wise'.
We can go back to Soln3 now. The loop order there is 'row-wise'. Lets change it
tic
Soln6 = ones(T, N);
for n = 1:N
for t = 1:T
Soln6(t, n) = 3*x(t, n)^2 + 2*x(t, n) - 1;
end
end
toc
Soln6 0.201661 seconds.
Better, but still very bad. Single loop - good. Double loop - bad. I guess MATLAB did some decent work on improving the performance of loops, but still the loop overhead is there. If you would have some heavier work inside, you would not notice. But since this computation is memory bandwidth bounded, you do see the loop overhead. And you will even more clearly see the overhead of calling Func1 there.
So what's up with arrayfun? No function inlinig there either, so a lot of overhead. But why so much worse than a double nested loop? Actually, the topic of using cellfun/arrayfun has been extensively discussed many times (e.g. here, here, here and here). These functions are simply slow, you can not use them for such fine-grain computations. You can use them for code brevity and fancy conversions between cells and arrays. But the function needs to be heavier than what you wrote:
tic
Soln7 = arrayfun(@(a)(3*x(:,a).^2 + 2*x(:,a) - 1), 1:N, 'UniformOutput', false);
toc
Soln7 0.016786 seconds.
Note that Soln7 is a cell now.. sometimes that is useful. Code performance is quite good now, and if you need cell as output, you do not need to convert your matrix after you have used the fully vectorized solution.
So why is arrayfun slower than a simple loop structure? Unfortunately, it is impossible for us to say for sure, since there is no source code available. You can only guess that since arrayfun is a general purpose function, which handles all kinds of different data structures and arguments, it is not necessarily very fast in simple cases, which you can directly express as loop nests. Where does the overhead come from we can not know. Could the overhead be avoided by a better implementation? Maybe not. But unfortunately the only thing we can do is study the performance to identify the cases, in which it works well, and those, where it doesn't.
Update Since the execution time of this test is short, to get reliable results I added now a loop around the tests:
for i=1:1000
% compute
end
Some times given below:
Soln5 8.192912 seconds.
Soln7 13.419675 seconds.
Oli 8.089113 seconds.
You see that the arrayfun is still bad, but at least not three orders of magnitude worse than the vectorized solution. On the other hand, a single loop with column-wise computations is as fast as the fully vectorized version... That was all done on a single CPU. Results for Soln5 and Soln7 do not change if I switch to 2 cores - In Soln5 I would have to use a parfor to get it parallelized. Forget about speedup... Soln7 does not run in parallel because arrayfun does not run in parallel. Olis vectorized version on the other hand:
Oli 5.508085 seconds.
That because!!!!
x = randn(T, N);
is not gpuarray type;
All you need to do is
x = randn(T, N,'gpuArray');