Matrix processing performance

A few days ago, fellow Matlab blogger Roy Fahn, well-respected in the Israeli Matlab community, posted an interesting article on his MATLAB with Fun blog (note the word-play). Since his article is in Hebrew, and the automated Google Translation is somewhat lacking, I thought of sharing Roy’s post here (with his permission of course), for the minority of the Matlab community which is not fluent in Hebrew…

Roy’s translated post: “Anyone who adds, detracts (from execution time)”

In the story of Eve and the serpent, the first woman told the serpent about the prohibition of eating from the Tree of Knowledge, adding to that prohibition a ban on touching the tree (something that God has not commanded). The snake used this inaccuracy in her words, showing her that one can touch the tree without fear, and therefore argued that the prohibition to eat its fruit is similarly not true. As a result, Eve was tempted to eat the fruit, and the rest is known. Jewish sages said of the imaginary prohibition which Eve has added, that this is an example where “Anyone who adds, [in effect] detracts”.

Recently I [Roy] came across an interesting phenomenon, that in MATLAB, adding elements to a vector on which an action is performed, does not degrade the execution time, but rather the reverse. Adding vector elements actually reduces execution time!

Here’s an example. Try to rank the following tic-toc segments from fastest to slowest:

x = rand(1000,1000);
 
% Segment #1
y = ones(1000,1000);
tic
for i=1:100
    y = x .* y;
end
toc
 
% Segment #2
y=ones(1000,1000);
tic
for i=1:100
    y(:,1:999) = x(:,1:999) .* y(:,1:999);
end
toc
 
% Segment #3
y=ones(1000,1000);
tic
for i=1:100
    y(1:999,:) = x(1:999,:) .* y(1:999,:);
end
toc

The first loop multiplies all the elements of the x and y matrices, and should therefore run longer than the other loops, which multiply matrices that are one row or one column smaller. However, in practice, the first loop was the fastest – just 0.25 seconds on my computer, whereas the second ran for 1.75 seconds, and the third – 6.65 seconds [YMMV].

Why is the first loop the fastest?

The subscription operation performed in each of the latter two loops is a wasteful action, and therefore in such cases I would suggest that you run your operation on the full matrix, and then get rid of the unnecessary row or column.

And why does the second loop run faster than the third?

This is related to the fact that MATLAB prefers operations on columns rather than rows. In the second loop, all the elements are multiplied except those in the last column, while in the third loop all the elements that have been extracted from all rows are multiplied, except for the last row.

In your work with MATLAB, have you encountered similar phenomena that are initially counter-intuitive, such as the example described above? If so, please post a comment below, or directly on Roy’s blog.

Is all of this undocumented? I really don’t know. But it is certainly unexpected and interesting…

Categories: Low risk of breaking in future versions, Stock Matlab function, Undocumented feature

Tags: , ,

Bookmark and SharePrint Print

14 Responses to Matrix processing performance

  1. I think this is really more an indication of how poorly MATLAB is written, than anything else. It seems that if you do subscripts on the right hand side, instead of optimizing that out by simply truncating the iteration internally, MATLAB actually creates an entirely new copy. In fact, MATLAB is so poorly optimized that if you replace your first segment with

    y = x(:,1:1000) .* y(:,1:1000);

    you will find that it runs just as slow as the second segment.

    Given that the MathWorks essentially has a monopoly on scientific numerical analysis software, I’d say it’s not very surprising that the core language implementation is of such low quality. I will grant you this, though: the fact that MATLAB, despite being claimed to have an optimizing JIT compiler, does something so stupid as to make an entirely new copy of an array when you only want to read a subset of it is definitely something the MathWorks doesn’t want to document…

  2. One thing to add: I originally thought maybe the problem was that the original example used in-place assignment, since y was on both sides. This isn’t the issue, and the results are the same (including a huge slowdown for replacing x with the “equivalent” x(:,1:1000)) if you use x for both variables on the right side.

  3. celine says:

    In the same subject, the function squeeze can actually be slower than reshape. Example below

    %%% evaluate the timing of 'squeeze'
    % create 3D tables
    apple=10*rand([100,300,400],'single');
    maxLoop=100000;
     
    % pick one line and one column
    tic
    for i=1:maxLoop  %this will take 3s
        core=squeeze(apple(3,25,:));
    end
    toc
     
    tic
    for i=1:maxLoop     %this will take 0.3s
        core2=apple(3,25,:);
        core2=reshape(core2,[1,400]);
    end
    toc
     
    tic
    for i=1:maxLoop       %this will take 0.3s
        core3=apple(3,25,:);
        core3=reshape(core3,[400,1]);
    end
    toc
  4. rami says:

    The reason the second loop is faster than the third loop is that matlab uses Column-Major Order :

    a = [1,2,3;4,5,6]

    is stored as 1,4,2,5,3,6.

    (Ref : http://en.wikipedia.org/wiki/Row-major_order#Column-major_order)

    As for the first one being faster than the second one, I dont know.

    Cheers

  5. Ajay Iyer says:

    Interestingly, here is a segment of code (#4) that is as fast as the #1.

    % Segment #4
    y=ones(1000,1000);
    tic
    x = x(:);
    y = y(:);
    for i=1:100
        y = x.*y ;
    end
    y = reshape(y,1000,1000);
    x = reshape(x,1000,1000);
    toc
  6. Thomas M. says:

    A little outside the subject, but I’ve been developing code using Simulink for xPC Target and the bottleneck in my algorithm is the large matrix calculation of A = M*N, where both M and N are (250,250). I am sampling at 0.001 and I am trying to find ways to optimize the speed. Is it possible to make this calculation faster? Also, why would this calculation take longer with xPC 5.0 compared with xPC 4.3 loaded on the same system?

    • @Thomas – You can always use a GPU or MEX to speed up the calculations, but try to think whether you really need to have all this data. Just reducing the size to (100,100) will improve performance by 2 orders of magnitude, not to mention reducing memory problems, thrashing (memory swaps) etc.

    • Thomas M. says:

      The matrix has to be that size. How can the GPU be leveraged on xPC Target? Also, what do you mean by using MEX to speed up calculations?

  7. Pingback: Waiting for asynchronous events | Undocumented Matlab

  8. Peter Gardman says:

    Dear Yair,
    I’m wondering about the fastest way of creating a structure. In my ignorance, I used to think that >>A.B = x would call A = struct(‘B’,x) so writing A = struct(‘B’,x) in an mfile would be much faster. To my surprise A.B =x is much faster:

    >> tic, A.B = 3; toc
    Elapsed time is 0.000003 seconds.

    >> tic, M = struct(‘B’,3); toc
    Elapsed time is 0.000297 seconds.

    Do you have any clue why is this?, or, if you have already talked about this in a post, could you please indicate it to me where? Thank you very much

    • @Peter – I suspect that this is due to the fact that A.B is directly interpreted by the Matlab interpreter, whereas struct(…) goes through a library routine (in libmx.dll on Windows).

      Note: when you place the profiled code in an m-file, rather then testing in the non-optimized Command Prompt, the speedup is only ~2 or 3, not 100 (not that a x2 speedup should be disparaged…).

      In a related matter, I compare alternatives for preallocating arrays of structs here.

    • Peter Gardman says:

      Thank you very much Yair

  9. Pingback: Python:Performance of row vs column operations in NumPy – IT Sprite

Leave a Reply

Your email address will not be published. Required fields are marked *