tract subsets, be sure you always perform the same actions on a list of increasing numbers. So, before you start to manipulate a list of 100 points, create a list of 100 integers (0, 1, 2, ..., 99) and make sure it gets mutilated in the exact same way as the pointlist.
Then, when all your points are modified, bring them all into the same list again and sort that list using the integer array as keys. This ought to put them back into the right order.
2) Reverse Engineering: since you know all your points are along well defined curves (lines in your case), you could project them all onto a line that spans the entire model. This will give you a list of curve parameters (floating point numbers). You can then sort your points once again, but this time using the parameters as keys. (See image: by sorting all the points using the curve parameters, you get the order in which they appear from left to right)
2b) If you need to do this thing on points which are in a grid (i.e. 2D sorting), you have to project onto a surface so you get uv parameters for every point. Then vastly multiply only the u (or only the v) values, since you want to give rows (or columns) a higher weighting. Finally add u and v together and this will give you another list of floating point numbers which can be used as a keys array in a sort operation.…
.0001, the functionality is been put into dedicated components (see this post for further details).
Different branches are always combined using Longest List logic. I'm unhappy about this as well, I need to give more control over how different branches are combined, but I haven't figured out yet how to expose such functionality without it being utterly incomprehensible to 99% of the users.
If you want to ignore the data inside the fourth branch, you'll need to remove that branch before the data goes into the Line component. It's easy to remove a specific branch, somewhat trickier to make this removal dependant on variables elsewhere in the network.
You can use the Split Tree component to achieve this either way. Using a fixed mask (like in the image below) may be sufficient.
The !3 means that any branch is allowed except when it has a 3 in that location. The [0-2] means that only branches which have a number in between and including 0 and 2 will be allowed.
--
David Rutten
david@mcneel.com
Poprad, Slovakia…
d" floor side).
Rails are obviously defined with slope adjustments at start/end: Imagine a rail ramp curve made via, say, 20 control points: at start p0 is not moved, p1 is moved half the step .... p19 is moved half the step*17 and p20 is moved the full distance. Thus we have what is called "slope adjustment" in our trade.
a myriad of options controls where the spaghetti starts (curve.PoitAt(userControllableT)) what is the continuity mode (sequential or steady[shown]) and what type of profile is used for the sweep.
…
that ... blah, blah) but each of those cores will be running at lower speeds because of the thermal restrictions.
For instance, a dual core may have base clock speeds of 3.5 GHz for each processor while a quad core processor may only run at 3.0GHz. Focusing to single core (on each of them) the dual core processor will be able to about fourteen percent faster than on the quad core. Thus, if you have a program/app that is only single threaded (99% of what's available for AEC puproses to be honest), the dual core processor is actually better. Then again, if you have something that can (?) use (??) all (???) four processors such as the notorious Nexus rendering engine (Modo/Microstation/AECOSim), then the quad core processor will actually be about seventy percent faster than that dual core processor.
But no AEC engineer worth his name cares about rendering stuff, he he.
AMD and Intel have introduced technologies that can dynamically increase the speed of a processor core to help offset these differences between the dual and quad/octa core products. For instance, Intel may have the quad core processor with a base clock speed be 3.0GHz but when only a single processor core is in use at full load, that processor core will be boosted up to 3.4GHz. This would then make the quad core processor just three percent slower than a dual core processor that runs at 3.5GHz.
In general and theoretically, a multiple core processor is a "better" choice but that does not necessarily mean that you will better overall performance.…
switch this talking off-line if you are interested to know the real reasons in depth.
What is the pro way? Well ... imagine objects (blobs et all) that are placed in 3d space by some per object policy whilst their "property" (bend,repulse) is user controlled on a per object basis. Then imagine variants of all that spaghetti yielded (the rays, that is) stored in parameters in order to do the obvious : take control of all your previous attempts (replace, remove, swap, reset etc etc).
Get a 10-- minute thingy (straight out of my head: NO checks OF ANY kind performed [bugs possible], just a grid that shoots rays and a single blob (a sphere) that does the job). Not even a decent random policy is applied in order to have some nice looking rays (not to mention their directions).
Now ... imagine any collection of breps distorting the ray chaos: i.e. a ray meets a blob > is distorted (or not) > then meets another > ... > blah, blah (plus some policy for killing rays heading to Sahara instead of Vienna - but that's elementary).
This requires at least 2 hours of coding to do it properly (+ the variants "management" C#).
But ... well ... it could be a good real-life case when Solaronix "sponge" type of U/V collectors could be available (rather soon) > I'll do it > the future > the glory > the cash > the polar bears.…
gain profits falls in the latter category.
The challenge here is to do the job (up to a point) using "anodyne" ways at the cost of providing slow, incomplete and quite inefficient solutions.
That said this specific vault case requires addressing 4 "classes" of problems (for instance: regions due to ccx events or alternatively circuits in graphs etc etc).
Back to business:
Creating a realistic "random" W truss of that type is one of the most challenging tasks in parametric adventures (in fact ... is the top dog by some miles). One of the many issues is an approach to manage "on-the-fly" clash situations by individually modifying nodes (without been sure that you can arrive to an all overall valid solution). Since one "path" tried may yield dead-end(s) this means keeping track of your corrective actions in a hierarchical manner and been able to follow a different "path". Another (obvious) issue is to use instance definitions for all the "components" thus achieving almost real-time response (try to manage 100K++ "solids" [sleeves, cones etc etc] to see what I mean) ... etc etc.
The big thing is: what are you going to tell to your instructors about the required code part? (that 99% mentioned) And if a "complete" solution is primarily based on "black boxes" could - in the instructor's eyes - your Master Thesis qualify as yours?
That said Vaults_V1 is achievable solely via components.…
ce attractors
3- Relation between mathematics and Form
4- Network surface and Paneling
5- Fabrication methods (slice3d, nesting, ...)
6- Structure and Architecture (Millipede)
7-Energy and form
8- Islamic patterns
9- Physics with kangaroo
…
ions are probably reflective of the prevailing humidity conditions (I just had a chat about this with my advisor, who incidentally also happens to be on the committee for LM-83).
The Tregenza sky patches considered in daylighting calculations don't do a good job of incorporating the correct size of the sun into calculations. In the figure below, the sun on the right is the one considered for calculations in Daysim. You can get a more accurate answer by considering a more discretized sky, however, I am not aware if that is possible with Daysim (and therefore HB) right now. Therefore, your direct sun calculations are likely to be off somewhat depending on how much of it there is(I'd say overestimated).
The calculations with humid sky, which are on account of the sky itself (and not the sun alone) are likely to be more relevant.
Regarding your questions about studying weathering effects with LB/HB, I have no idea as that is something that I haven't looked into before. I am sure someone else on this list has a more informed opinion on this issue than I do.
Your project, and your approach to it, seems really interesting and I am glad to be having this discussion :).
Sarith
…
try now to integrate Geco in an interdisciplinary architectural engineering studio: hoping we can show you some nice applications of your tool, I'll keep you update and sending now details by e-mail. Here the file (very welcome to be shared). It most probably contais trivial errors by me, thanks for helping and giving some tip! Gr. Michela
FILE:
Ok, right, I see the outputs update correctly. Origin of problems must be in some different mistake I do:
- Incident radiation: I am not sure I understand what is going on: why I get so many 'not a number' ? (The Galapagos report is full of NaNs).
Bio-Diversity: 0.887 Genome[0], Fitness=NaN, Genes [89% · 44%] { Record: Too many fitness values supplied } ...
Genome[7], Fitness=NaN, Genes [74%] { Record: No fitness value was supplied } ....
Genome[9], Fitness=NaN, Genes [37% · 11%] { Record: Genome was mutated to avoid collision Record: Too many fitness values supplied }
- Daylight calculations: the geometry accumulates withouth deleting the previous models. As a consequance, results almost do not change after few varations (so, outputs get updated but do not vary). In current daylight definition: the first object being imported is the one where the grid has to fit; its setting makes it cancelling all the other objects during import. All the others, do not delete anything when imported. When running loops (manual or GA) that vary parameters, the entire geometry do not get cancelled - so I guess the loop does not pass back by the cancelling step, but imports only the geometry which has been varied by the parameters using the setting of that import component only? I will then try again by changing the order of the operations, but if you have specfic tips, let me know.
THANKS!
…