If we're to speak of intuition for new developers who've never used a
programming language, who are falling back to what they know about
mathematics, then quite literally a decimal point _is_ about division by
ten.
I don't think this necessarily follows; the issue here is that the
constructor isn't explicit enough that it is simply lopping off the
fractional part. My own experience of maths as taught in school, to go from
a decimal to an integer I would expect to round,
You would also expect that 3 / 4 in integer math gives you 1. With integer
division, however, 3 / 4 == 0. By definition the decimal point separates an
integer from a fractional part, so the behaviors are inextricably linked.
To test this out in practice, I asked the first person with no programming
experience I just encountered today.
I said: "Let me teach you one fact about integers in a programming
language. When two integers are divided, the integer result has the
fractional part discarded; for example, 3/4 computes to 0. What would you
expect to be the result of converting 0.75 to an integer?"
He answered immediately: "I would have expected that 3/4 gives you 1, but
since 3/4 gives you 0, I'd expect 0.75 to convert to 0."
These are two different case; Int(3) / Int(4) is a division of integer
values with an integer result, there's no intermediate floating point value
that needs to be coerced back into an Int. The issue here is converse of a
Float/Double to an Integer, it's a different operation.
3.0 / 4.0 = 0.75 is a property of Float
3 / 4 = 0 is a property of Int
What's under discussion here is conversion between the two, they're not
really comparable cases.
We're not disagreeing here. Or at least, I'm not disagreeing with you. Of
course, integer division is not the same operation as conversion of Float
to Int. However, division and decimals are related concepts, obviously. I
don't think you dispute that.
How integer types behave under division and how they behave when converting
from a decimal value are therefore inevitably related. And as my
non-programming colleague demonstrated today, new users will use their
knowledge regarding the behavior of one operation to shape expectations
regarding the behavior of the other.
Discarding the fractional part of a floating point value is a bit pattern
operation only in the sense that any operation on any data is a bit pattern
operation. It is clearly not, however, an operation truncating a bit
pattern.
as the conversion is simplistically taking the significand, dropping it
into an Int then shifting by the exponent.
That's not correct. If you shift the significand or the significand bit
pattern of pi by its exponent, you don't get 3.
I think you're misunderstanding me. If you think of it in base 10 terms
1.2345 is equivalent to 12345 x 10-4; when you convert that to an Int it
effectively becomes 12345 shifted four places to the right, leaving you
with 1. In that sense it's a truncation of the bit-pattern as you're
chopping part of it off, or at the very least are manipulating it.
No. It is of course true that the most significant bit of the binary
representation of 12345 is 1. However:
12345 >> 1 == 6172
12345 >> 2 == 3086
12345 >> 3 == 1543
12345 >> 4 == 771
12345 >> 5 == 385
etc.
That is to say, you can't get 1234, 123, or 12 from truncating 12345. There
is no sense in which converting 1234.5, 123.45, or 12.345 to an integer
involves truncating the bit pattern of 12345. You are not in that case
performing recursive integer division by 2, but rather recursive integer
division by 10, which goes back to how decimals and division are
inextricably related.
Regardless it's also very literally a truncation since you're specifically
truncating any fraction part, it's simply the most correct term to use;
frankly I find restricting that to bit-pattern truncation to be entirely
arbitrary and unhelpful.
It is important and not at all arbitrary. Binary integers model two things
simultaneously: an integral value and a sequence of bits. Much care was
placed during the design and review of the revised integer protocols in
making sure that the names of operations that view integers as integral
values are distinguished from those that view integers as sequences of
bits. It was accepted that "truncating" and "extending" would be applied to
operations on bit patterns only, which is why it was OK to shorten the
label from `truncatingBitPattern` to `truncating` (later renamed
`extendingOrTruncating`, for other fairly obvious reasons). By analogy, the
expected value for a hypothetical `Int32(truncating: 42.0 as Double)` would
be 252867936, which is of questionable usefulness. It would, however, be
confusing and unhelpful to use the same word to describe an operation on
the represented real or integral value which is now used only for a very
different operation on a sequence of bits.
The types involved should make it clear whether the value is being made
narrower or not. Int64 -> Int32 is a form of truncation, but so to is Float
-> Int; in both cases the target can't represent all values of the source,
so something will be lost.
func init(rounding:Float, _ strategy: FloatingPointRoundingRule) { … }
Again, here, as an addition to the API, this fails the six criteria of
Ben Cohen, as it is strictly duplicative of `T(value.rounded(strategy))`.
Maybe, but init(rounding:) is explicit that something is being done to
the value, at which point there's no obvious harm in clarifying what (or
allowing full freedom). While avoiding redundancy is good as a general
rule, it doesn't mean there can't be any at all if there's some benefit to
it; in this case clarity of exactly what kind of rounding is taking place
to the Float/Double value.
The bar for adding new API to the standard library is *far* higher than
"some benefit"; `Int(value.rounded(.up))` is the approved spelling for
which you are proposing a second spelling that does the same thing.
The main benefit is that the constructor I proposed would actually require
the developer to do this, what you're showing is entirely optional; i.e-
any value can be passed without consideration of the rounding that is
occurring, or that it may not be as desired. With a label the constructor
at least would remind the developer that rounding is occurring (i.e- the
value may not be as passed).
I do not think any users expect 0.75 to be represented exactly as an
integer; that's not at issue here. The question is, which users see
`Int(0.75)` and think, "this must mean that 0.75 is rounded up to 1"? My
answer, from prior experiences teaching beginners, is that the subset of
users who make this mistake (or a similar one in other languages) largely
overlaps the subset of users who see `3 / 4` and think "this must evaluate
to 1."
Going further and requiring them to provide a rounding strategy would also
force them to consider what method of rounding should actually be used,
eliminating any confusion entirely. What you're demonstrating there does
not provide any of these protections against mistakes, as you can omit the
rounding operation without any warning, and end up with a value you didn't
expect.
It is, of course, true that a user who does not read the documentation and
expects a function that does A instead to do B will use that function
incorrectly. The question is whether it is reasonably common for users to
make the incorrect assumption, and whether such incorrectness ought to be
accommodated by a breaking change to the language. Here I am arguing that
users who are aware of integer division are unlikely to make the incorrect
assumption; they will at minimum look up what the actual behavior is, and
based on my teaching experience and today's mini-experiment, they are
likely actually to expect the existing behavior. (And, of course, I am
arguing that users who are unaware of integer division have a much more
serious gap in knowledge that is the primary issue, not fixable by tweaking
the name of integer initializers.)
A secondary benefit is that any rounding that does take place can do so
within the integer type itself, potentially eliminating a Float to Float
rounding followed by truncation; i.e- since rounding towards zero is the
same as truncation it can optimise away entirely.
Renaming the `Int.init(_: Float)` initializer, by definition, cannot
recover any performance benefits. I'm not aware of any optimizations of
other rounding modes that can make such an initializer faster than what is
currently possible.
···
On Mon, May 22, 2017 at 5:21 PM, Haravikk <swift-evolution@haravikk.me> wrote:
On 22 May 2017, at 21:16, Xiaodi Wu <xiaodi.wu@gmail.com> wrote:
On Mon, May 22, 2017 at 10:39 Haravikk <swift-evolution@haravikk.me> > wrote:
On 22 May 2017, at 15:51, Xiaodi Wu <xiaodi.wu@gmail.com> wrote: