Rhinoceros y Grasshopper los participantes puedan desarrollar mediante un flujo de trabajo digital las técnicas de Diseño Paramétrico, mediante métodos algorítmicos generativos, para su implementación en procesos de diseño y representación de los mismos, a través de:
Proliferación en topologías complejas mediante Definiciones visualizados Estrategias de Mapeo y Análisis evolutivos en Tiempo Real Técnicas avanzadas de modelación 3D e imágenes foto realistas Vinculaciones entre criterios Predefinidos paramétricamente y Morfologías resultantes
Detalles: ∙ Instructores / Matías Casciotta y Salomón Escobar / Integrantes Ha-11 ∙ Todos los niveles de experiencias son bienvenidos ∙ Recursos: Interfaces Rhino + Grasshopper / Manuales PDF / Definiciones GH Preparadas ∙ Los participantes deberán traer su propia laptop con todo el software y actualizaciones (originales o versiones de demostración) previamente instaladas. (Se fijara una fecha unos días antes para revisar que todos los equipos estén en orden y listos para trabajar).
Fecha y Lugar ∙ 3 semanas a partir del 15 de Marzo del 2012 - Lunes y Jueves de 19:00hs a 21:00hs - ∙ El taller se dictara dentro del ámbito de cursos del Arquitecto Alejandro Correa.
INFORMES: Ha11 en Facebook ha.11.arquitectos@gmail.com…
supplied _values of _keys" notice.I tried running the "OSM 3D" component first with groundTerrain_ input. As I did not get the upper notice message, I closed down the whole Rhino so that I cut the waiting time. Then I tried running it without the groundTerrain_ input, and in some 15 minutes I got the following buildings:
I think I may understand what was causing the problem: when one takes large radii, it covers large areas, and with this area comes large number of information (keys and values). You can get hundreds of keys (or thousands). What can happen is that: these hundreds of keys, can exceed shapefile's capacity to story keys. So basically in case of radius 750 meters your "height" or "buildings:levels" keys somehow slipped beyond this allowable capacity. In case of 800 meters they were somehow allowed to enter (a bit bad term sorry) before the allowable capacity is reached. This depends on the number of keys named with letters which precede the "h" and "b".The best way to solve this issue is to know which data do you actually need, and use the "OSM Keys" component to generate the list of needed keys. In this way, only those keys that you need will be used, others will be disregarded.You do not even have to use the "OSM Keys" component if you know which specific keys you exactly need. Check the attached file below. I grouped the "OSM Keys" solution as "a" and a custom defined list of keys as "b".
2) The component running time might now be cut with picked "requiredKeys_" input I mentioned at the end the previous 1) part.
3) "OSM 3D" component's "randomHeightRange_" input is suppose to do exactly that: to randomly create 3d buildings (or 3d trees) when there are no valid "height" or "buildings:levels" tags.I have just changed one line the "OSM shapes" component code.I wonder if it would make any problem on your PC.Please let me know if LocationGrabber03_Gismo2.gh file works.…
Added by djordje to Gismo at 2:34pm on February 11, 2017
400m swatch from a point somewhere in Catalunya.
The three APIs used were the following:
Google Elevations API
Mapquest Open Elevation Service API
Geonames SRTM3 API
I also tested the USGS Elevation Service, but I was looking for API which allowed me to query globally.
Here are the results (441 locations queried):
As a side note, Grasshopper reports the requests for data came in at*:
1.9s for Google Elevations API
3.5s for Geonames
413ms for Mapquest
*this is not only measuring the request, but also has to take into account the request throttling due to the various API limitations.
As you can see, there is quite a difference in the data, especially when looking at what Google returns. It is pretty clear that Mapquest and Geonames use very similar data coming from the SRTM3 dataset. This dataset is at 3 arc-seconds (appx 90m) for most of the globe (up to 60ºN and 56ºS). The resolution is 1 arc-second for the United States. Google reportedly uses hundreds of data sources to achieve a finer resolution, though this comes at a cost. Geonames and Mapquest put a limit of how many locations you can query at one time, with no limit per day (that I could find). Google puts a limit of 2500 requests per day, with each request having up to 512 locations, or a total of 25,000 locations.
The comparison was made possible by some of the little utility components which are included in gHowl, namely the XYZ->GEO component which translates points in Rhino/Grasshopper to WSG84 coordinates. …
ilion.
Then i sketched the outline curves in rhino with a few control points. The building is symetric so i only draw one side. But i'm not sure what is better for a voroni. a sharp or a soft surface? Or dose i need points?
So i have some questions:
1. how can i loft the curves correctly? My problem is that if i divide my curves for more control points, grasshopper automatically change my curve. thats ok but than i've the problem with a short curve, which fit bevor with the large one, but after the devision it can't connect.
So i tryed to duplicate the long curve and split it but with the shatter battery it dosen't work. It always cut the curve somewhere.
2. my next problem is, the curves in rhino should be my main construction, which is always visible. so i decided to offset the curves that i got a colum. but i don't know how to orient the offset curves in the xyz axis.
3. hopefully if i have the surfaces, how can i build a voroni which is offsetet, and has maybe some different thicknesses? :D
Would be really great if s.o. can help me. I tried a lot but not every thing is simple.
Sorry for my bad english.
Thx max
Here are my files:
FCP_MAX_GH_konstruktion_1.3dm
FCP_MAX_GH_konstruktion_1.gh
…
w how. Thanks for that. Now I do have some questions.
1. I am using the area weight tool. I am first calculating the volume of the form. I then multiply that value by it's density. So for concrete I am using 2400 kg/m^3 x volume. I then divided that number by the area of the membrane that is supporting the mass. This gives me my area weight. It seems to be working well but I want to verify that this is the correct workflow. I also want to verify that gravity would be turned off since I am thinking it is already calculated within the weight component.
2. I am finding that the new triangular element tool works much better than trying to use EA/L as input for the springs from mesh. Even when I set the timestep, subiteration, and drag I still have issues with getting very stiff materials to work. On the new finite elements tool I wanted to verify that E was in pascals. I also wanted to ask if I use imperial units can psi be entered. Now from what I am seeing the materials are deforming more than expected and to get less deformation and stretch in the mesh area I am finding the E value needs to be increased more than the true material values. Often I am raising E by a multiple of 10 or 100.
I am going to describe my problem and I will gladly share the definition if you'd prefer looking it over but basically I have an inflated membrane at a certain pressure made of a particular material. I then have a certain volume of concrete on top of the inflated membrane. My goal is to review the displacements as the concrete is applied over the membrane and find the proper pressures to apply to keep it free from deformation. I am including a picture from a project that we used kangaroo on and attempted to deal with such issues. It was a class sponsored by Cloud9 architecture held at Art Center College of Design where I was one of the instructors. Hopefully this illustrates the problem. To summarize any example file that shows the best way to implement real material properties and unit based forces would be a helpful reference and would be greatly appreciated.
…
trying to develop it for my own project.
http://www.grasshopper3d.com/forum/topics/shortest-walk-tapered-branching-script?xg_source=activity&id=2985220%3ATopic%3A1450323&page=2#comments
On this page, he shared few 3D coral difinitions and especially interested in first and second one.
First one( bunny like 3D coral) - posted on February 2, 2016 at 9:43pm
Second one( sofa like 3D coral) - posted on February 6, 2016 at 3:16am
I followed these instructions, succeeded to build Tetgen, placed the built files in C drive directory and tried to run the definition. Then some WindowsError came out as follows which I don't know how to fix.
My working environment is;
OS is Windows 7 Ultimate 64 bit.
Rhino is version 5, 64 bit.
Grashoppper is version 0.9.0076, the latest version at this moment.
It would be great if I can have some help advice / comment.
I appreciate for your attention.
…
printers.
How I want to communicate this: The depth of transparent cubes is relative to the brightness of a picture (low depth = bright, high depth = dark). Then I assign each cube as red or blue depending on the RGB values of the cube column's corresponding pixel - this is where I'm stuck.
What I've done: I have one image sampler containing a greyscale version of my image which is outputting the brightness measurements. This made into lines, which are divided to create the points from which the cubes are created. (I have had to invert the image in photoshop as brightness gives black a low value when I need a high one, and vice versa)
What I want to do next: In the second image sampler I have an image which has a Red to Blue gradient applied to it. I want to group my cubes into reds and blues depending on the colour values in this image (so they could eventually be saved as a "blue" and "red" stl to be 3D printed).
So columns that correspond to a blue part of the image will contain a completely blue stack of cubes, and the same with red. But where there's a combination of blue and red values I need a combination of blue and red cubes mixed together. I was hoping to do this by turning the RGB values into some kind of ratio that will help assign each cube a group but I'm struggling.
Would love any thoughts on resolving my problem, even if it's only for part of it! This was quite hard to explain so let me know if there's anything that needs clarifying.
Thanks…
mment%3A1637953
First of all, the invalid Rhino license as seen previously has been removed, and the correct educational license we have is re-installed for this test.
The re-appearing issue is that RAM usage spikes once GH is open in Rhino. It seems that this happens when a series of large GH project files incrementally saved are stored in the same folder. Moving those previously saved large project files to a new folder seems to be able to solve this issue.
The images below explains the issue and the hypothetical solution:
1. A series of GH files were incrementally saved in the same folder previously, and the last few GH files are the ones opened most recently:
2. The total RAM usage is at the normal 5GB level once Rhino is open:
3. Once GH is open, the RAM usage spikes, and the it becomes very slow to maneuver the GH window before even opening any one of those GH files:
4. Once GH and Rhino are closed, the RAM usage drop to the previous level before the GH interface was open:
5. Now, all the incrementally saved GH files are moved to a new folder "wip" except the last one, i.e. for the last GH file, there is no other previous GH files in the same location:
6. Now, if we open GH, there is no sudden increase of RAM usage, and the 3x3 thumbnails on the GH canvas shows "missing" as those previously opened GH files are no longer in the same location as they were before:
I understand that David mentioned that the thumbnails for previously opened GH files on GH canvas will not take much RAM. Nevertheless, I'm still not sure what is causing the increase of RAM usage and slowdown of GH interface. Relocating the large project files previously saved in the same folder as the current GH file seems to be able to make this issue go away, for unknown reason ...
Appreciate if anybody experiencing similar issue can help to check if this solution works.
Thank you.
…