[FRIAM] edgelords
glen
gepropella at gmail.com
Thu Jun 9 13:07:54 EDT 2022
Agreed. A colleague and I have a disagreement about how to specify a container. Because he works almost exclusively in Python, he uses python-based specifications. That tracks with the data scientist in charge of the locked down portal who has 1 container for R work and anther for Python, both wired to reach out to host OS volumes even for "local" data. Because my stack is a gooey mess of tools including things like sed, awk, and node, as well as using both Python and R at the same time, I tend to build my containers as if they're full blown VMs. I go inside, intra-container. Their containers are fine-grained and the gooey bits are inter-container. I suppose if/when there's a chance to parallelize the workflows, their way is better. But when it's mostly serial, my way maintains repeatability better. Plus it allows me to organically arrive at automation through upstream manual repetition, rather than somewhat idealistic planning with error correcting iterations.
On 6/9/22 09:04, Marcus Daniels wrote:
> A reason to run locally is to know what you have and to be able to control everything. My sense is that people waste a lot of time maintaining containers.
>
>> On Jun 9, 2022, at 6:50 AM, glen <gepropella at gmail.com> wrote:
>>
>> The End of Localhost
>> https://dx.tips/the-end-of-localhost#heading-the-potential-of-edge-compute
>>
>> On the tails of the Get off my lawn! AOL thread, that localhost article reminded me of Firefox's new tool:
>>
>> https://addons.mozilla.org/en-US/firefox/addon/firefox-translations/
>>
>> I don't yet understand how it works. But assuming it's true, I like the idea that the translator robot runs on localhost. But it also invokes 2 problems I currently have: 1) coworkers who won't share their premature/broken works in progress and 2) the opacity of computation that happens elsewhere. If you read the Hacker News thread <https://news.ycombinator.com/item?id=31669762>, you see lots of yapping about "developers" and front-end stuff, not understanding back-end stuff, yaddayadda. And that's fine; gatekeepers are everywhere. But there are serious "openness" issues with relying on compute elsewhere. And it's not merely supply chain problems like what version are they running back there. One data portal my clients want/expect me to use prevents any traffic in or out, for data privacy reasons. But many of the workflows we use to knead data call out to online APIs, in my case so that you "don't have to worry about" what version of whatever lies on the other side. So, obviously, I have to convert all the outreach to localhost, either with simulated servers or installing large blocks into the container and refactoring network calls into local calls. That bloats my container, of course, slowing the development process. Well-simulated data becomes important so I can tighten the dev loop on localhost before sending the bloated container to the portal to test on real data.
>>
>> I'm no longer sure where I'm going with this. Sorry. Were I intelligent, I'd delete my commentary and just send along the links. Maybe SteveS has finally infected me. 8^D
--
ꙮ Mɥǝu ǝlǝdɥɐuʇs ɟᴉƃɥʇ' ʇɥǝ ƃɹɐss snɟɟǝɹs˙ ꙮ
More information about the Friam
mailing list