[Noisebridge-discuss] A test of Excellence
rachel lyra hospodar
rachelyra at gmail.com
Thu Oct 18 20:58:08 UTC 2012
On 10/13/2012 4:32 PM, Martin Bogomolni wrote:
>> This is an interesting thought experiment.
>
> It started as a thought experiment some months ago. It's progressed
> from there.
i describe it as such intentionally. Not because i have missed the fact
that your experiment has progressed on to reality, but because i believe
there are so many other unnacounted-for variables that it is in fact
still in the design stage. The reality that you are pursuing is too
flawed to be valid. For me the experiment is still in the "is this
viable, and what can we determine" stage. You'll forgive me for working
to reach my own conclusions in that regard.
>
>> How are you going to correct for different quantities of users in the various spaces?
>
> I chose the three spaces because they have approximately the same
> number of users, although the demographic of those users is obviously
> different. I actually did a population count over the last few
> months by coming on different days, and different times, and counting
> how many people were there at different times.
>
>> We don't even count usage statistics but maybe there is some way to build some data through observation and random sampling.
>
> Bingo :)
I'd like to see this data please. how many different times? there are
lots of external factors on space usage (eg university and holiday
schedules, or local events like maker faires) so you'll excuse me if i
don't just trust you that your sampling has been broad enough to be
relevant. Additionally, the demographic differences that you gloss over
are statistically influential and should be acknowledged even if they
cannot be measured directly.
>
>> What about correcting for the relative value of other objects surrounding the object in question? EG, in a space full of nicer things, the same identical object becomes less attractive to a potential thief because other juicier objects abound.
>
> By choosing five objects of differing values (social, intrinsic, tool,
> percieved etc), I'm doing my best to make them more-or-less match
> what's in the environment. i.e. make them all equally tempting.
>
Mmm, i don't know if i accurately conveyed the concern - if the groups
of objects are the same in value as each other, they will still have
different relative values in relation to the rest of their hackerspace
environments, since the different hackerspaces will have different
concentrations of objects of value, and additionally those objects of
value will have different relative values. Unless you have already
constructed this metric and have placed objects in the spaces that have
values that are equalized across the relative value of the other objects
in each space. In which case, please share your data.
Additionally, we have to note and account for the influence on relative
value that is created by proximity to the black market. At 16th and
mission, 2 blocks from NB, is a prime location for buying and selling
stolen goods, thus increasing the likelihood that a thief will be able
to quickly divest of the stolen goods, thus perhaps increasing the
likelihood they will be stolen. if you cannot measure a factor that
does not mean you can ignore it. sometimes you can note it and assign
it an approximate value.
>> That same principle also needs to be related, in a more complex way, to the neighborhood surrounding the space. Something tells me the relative crime indices of the neighborhoods vary, from the heart of SF Mission to Santa Cruz.
>
> Cultures and societies are self-selecting. All three locations have
> varying degrees of access control, and social selective pressure.
> Students at UCSC would find the objects as desirable as people
> wandering into the Dallas Makerspace.
If there are a thousand laptops always present in one space, and you add
one in a corner with a 'don't hack me' note, that is different than
doing the same thing where there are only ten laptops always present.
Please clarify.
This is at the heart of what
> I'm testing for, and all three handle social and physical access
> control slightly differently, but at the core .. it comes down to
> physical access control (door and key) and social pressure (are you
> supposed to be here?).
>
So let me try to understand. The heart of what you are testing for is
the persistence of similar objects in their respective locations. You
believe that this will somehow convey information about the relative
success of three complex access control schema in three different
environments?
This is a hypothesis with which i disagree.
>> Are you going to have a variable in your equation for the ineffable benefits various differences between the spaces have in other ways, besides perhaps influencing this single metric?
>
> Nope, because ineffability is ineffability. There is certainly
> something to be said if, for example, all five objects go missing at
> one place, but not in another. Or perhaps that none go missing, but
> never get put into the "correct" place, or perhaps that they will be
> used or abused. That will come out in post-analysis.
>
>> I am not certain that "whether my shiny expensive shit got stolen" is the only metric of mean honestly level in a group.
>
> Of course not! But it's one metric. Also, the relative value of
> "shiny" has been taken into account with the items. They are
> valuable, but not ridiculously or obviously so.
The calibration of this metric, of what is valuable, varies across the
different spaces.
It's entirely
> possible, that to combat the Hawthorne Effect, that I have another
> related experiment in progress but not announced in general. I have
> -already- run one such series without an announcement, and now I'm
> comparing the differences.
Overall I find this approach to social dynamics to be paternalistic and
relatively content-free. There is a whole lot of
trust-my-statistical-model handwaving. Any one person is going to have
blind spots, and especially in relation to geeks and social dynamics i
have a hard time with a statistical model around complex social
interactions, constructed by a single individual.
>
>> How likely you are to have to tackle a sticky-finger sketchbomb also needs to be balanced out with how likely you are to encounter a visiting troupe of foreign journalists who have come all the way to noisebridge seeking to learn more about TECHNOLOGY. from US. because we ROCK at technology and also at sharing, even when it is inconvenient.
>
> Indeed... and if sticky-fingered journalists walk away with something,
> it still goes back directly to the core of the experiment. How the
> group interacts with guests is just as important as how the group
> treats itself internally.
>
Let me know if you design an object-permanence experiment that purports
to measure statistically the way we interact with our guests, Mr. Scientish!
>> Not that I don't think we should be able to have nice things, but i think this approach is so reductionist as to be incapable of producing relevant data.
>
> I politely, and respectfully, disagree. I am testing, and have
> already tested, some of these issues in isolation from one another,
> keeping away from complexity on purpose.
A mechanical engineer who functions this way makes structures that
crumple. You must account for the variables that you aren't isolating
for, alongside the one you are. The way to create a simple, testable
situation is not to ignore all the variables you cannot quantify, or
assume they cancel each other out. Instead you must go through the
systems identifying all the variables that you can, and assigning them
approximate values or finding ways to turn them static (eg by bracing
every x interval, or renegotiating employment contracts). These
approximations are refined as you make each subsequent planning or
assessment pass through the system. This is how to construct a budget,
or a mathematical model of a static or dynamic system. I haven't
studied sociology or anything but something tells me human variables
matter in human dynamics as much as the variables that make up fluid
dynamics.
It's also a reason that I've
> been so quiet (when I'm normally rather gregarious) at noisebridge the
> last few months.
>
>> If you are interested in increasing the complexity of your model, I
>> have suggested some variables that need to be accounted for. I'd be happy
>> to give feedback on the ways that you are thinking of incorporating them.
>
> I'm always happy to collaborate and share ideas! That's the nature
> of being a hacker, after all. Although I should note that you're
> also part of the experiment group, and if you help design the
> experimental series, you can influence and thus invalidate the study.
> As long as you're unaware of the detail, though, the effect is
> negligible.
again, i assert that your study's design is still so rudimentary as to
be worthless. If you are interested in me, personally, believing that
what you generate even IS data, you can start by increasing the
complexity of your model, if needed redesigning your experimental
approach so that it can be discussed publicly.
R.
>
> -M
More information about the Noisebridge-discuss
mailing list