Revision as of 12:39, 8 November 2024 editMasataka Ohta (talk | contribs)25 edits →Binary Digit: ReplyTag: Reply← Previous edit | Revision as of 12:59, 8 November 2024 edit undoMrOllie (talk | contribs)Extended confirmed users, Pending changes reviewers, Rollbackers236,137 edits →Revisions to lead: new sectionTag: New topicNext edit → | ||
Line 371: | Line 371: | ||
I will add a new section for etymology referencing Shannon's original paper to have defined "bit" with two meanings (as a unit and as a symbol) and clarifies such misconceptions as a bit, as a unit, could have a state represented by 0/1, even though it's bit as a storage/symbol which has state to store/encode bit values. --] (]) 05:43, 12 April 2023 (UTC) | I will add a new section for etymology referencing Shannon's original paper to have defined "bit" with two meanings (as a unit and as a symbol) and clarifies such misconceptions as a bit, as a unit, could have a state represented by 0/1, even though it's bit as a storage/symbol which has state to store/encode bit values. --] (]) 05:43, 12 April 2023 (UTC) | ||
== Revisions to lead == | |||
This new addtion to the lead is confusing and poorly explained, and it seems like it relies on a paper written by the editor making the change - (see ], this is a problem in itself). If you are trying to mention the difference between a bit as used in digital communication and a ], so it would be redundant as well. ] (]) 12:59, 8 November 2024 (UTC) |
Revision as of 12:59, 8 November 2024
This level-4 vital article is rated C-class on Misplaced Pages's content assessment scale. It is of interest to the following WikiProjects: | |||||||||||
|
Binary Digit
Binary digit? A digit is the name given to a number used in the positional value notation when using base 10 --18.239.6.217 (talk) 03:56, 30 May 2008 (UTC).
- If that is true, you should probably correct the article Numerical digit before you do this one. - Soulkeeper (talk) 10:30, 10 January 2013 (UTC)
- Talk about picking nits. Why not just agree that a bit is defined as a binary digit and not worry about origins. Does it really matter if it was an awkward contraction ? Or that the usage may have actually originated in the subdivision of a Spanish gold coin? Whatever its origin, I believe that b for bit and B for byte works well and is commonly accepted as the standard notation.
- JamThi (talk) 14:00, 15 November 2013 (UTC)
- > Binary digit? A digit is the name given to a number used in the positional value notation when using base 10
- As I wrote in my IJANA paper:
- > Obviously, a “digit” means a decimal digit, a character of
- “0”, “1”, “2”, “3”, “4”, “5”, “6”, “7”, “8” or “9” ...
- > Similarly, the original meaning of a “binary digit” is a
- character of “0” or “1”, which may be represented by a
- binary symbol or storage cell.
- you are arguing against not me but the established and well defined concept of binary
- digit, which is fully recognized by Shannon.
- As I am already a author of several IEEE papers and RFCs, I chose IJANA for my new paper merely
- because no article processing charge is necessary. Masataka Ohta (talk) 12:39, 8 November 2024 (UTC)
It is stated that decimal forms of the SI factors (10^2, 10^4..) are used when related to telecommunications. Thus we can agree that 56 kilobits per second essentially means 56 000 bits per second. That thus gives us 7 000 bytes per second. However, when we carry on to convert to kilobytes per second, should we not use the binary form of the prefix, i.e. 7 000 bps = 7 000 / 1024 = 6.836 kilobytes per second?
We should not. When talking about telecommunications, use powers of ten always. -(struck by)Splash 19:49, Jun 15, 2005 (UTC)
- Regrettably, the orginal commentor was right. Both the binary forms of the SI factors and the standard ones are used. Bits are always measured in standard SI units, e.g. "10 = kilo", and bytes always on the "2 = kilo" etc scale. So 56kbits is 56000 bits, 7000 bytes and 6.84kbytes. See the binary prefix page. Incidentally, this is the source of much confusion (in the UK) at least, about why people's 1Mbit connection only yields 125kB/sec - the B used by a certain browser is in bytes, not bits. -Splash 19:49, Jun 15, 2005 (UTC)
- The "binary prefix page" does not support your contention that bytes on comm lines are "always" expressed using binary prefixes. The standard in telecom engineering is decimal prefixes. Some software does report many things using binary prefixes that it shouldn't - for example, Windows is infamous for reporting hard drive sizes using them, even though hard drives are not sold that way - but do not assume that this is any sort of standard. There is no conflict between "1 Mb/s" and "125 kB/s", not with eight bits to the byte. Note that browsers see only the application-layer data, not the IP and TCP or UDP headers, an issue more akin to file system overhead than anything about prefixes. Jeh (talk) 01:14, 25 March 2017 (UTC)
- We should use the IEC prefixes whenever it is confusing like this. — Omegatron 18:27, August 27, 2005 (UTC)
- I see that it has finally been resolved! =) - original poster
- I am seeing more examples where 1024 bytes are being displayed as kiB.
- Manufacturers always display the capacity of their memory or storage capacity using the largest numbers and thus prefer the 1000 bytes = kB format. I have seen some instances where 1024 bytes have been displayed as kB instead of kib. It can be very confusing if the new standard is not used.
- Transmission speeds are not a simple calculation of how many data bits are being sent. The header data must be factored in as well. Do you know if there are any start or stop bits included in each byte during transmission.
- JamThi (talk) 14:00, 15 November 2013 (UTC)
- Re "Manufacturers always display the capacity of their memory or storage capacity using the largest numbers" ... sigh... not that canard again. If that is true, then why is semiconductor RAM always quoted using binary prefixes?
- The fact (as described at Binary prefix) is that in the earliest days, such stats were often cited using no prefixes at all. e.g. the IBM 350 RAMAC, the very first hard drive on the market, was quoted as storing "5 million characters". (With 100 recording surfaces, 100 tracks per surface, and 500 characters per track there was nothing about it that leaned toward a power of 1024.) Main memory started being quoted using powers of 1024 simply because with a binary-addressed computer, memory naturally expanded by powers of two (unlike in hard drives). On the other hand there is nothing about communications lines that "prefers" powers of two. Jeh (talk) 01:14, 25 March 2017 (UTC)
bits
isn't a bit also 12.5 cents? as in two bits equating a quarter? as in a shave and a haircut, two bits? if so, that needs to be expressed on this article, and the derivation should be explained to. thank! Kingturtle 23:00, 24 Apr 2005 (UTC)
- This is an example of where the Spanish gold coin subdivisions was used to subdivide the dollar. There is no 'one bit' of a dollar. Two eights are a quarter which is the other notation used for a twenty-five cent piece. When half dollar coins were popular they were sometimes referred to as four bits. I never heard any other usages in relation to the dollar.
- JamThi (talk) 14:05, 15 November 2013 (UTC)
Absolutely. I agree that the article should make some observation that it makes sense linguistically to shorten Binary Digit to bit because of the history in English of a bit representing one eighth of a whole, as Binary Digits represent one eighth of a byte. Jlavezzo 16:16, 15 February 2007 (UTC)
Bit, trit, dontcare, what?
Two is to bit as three is to trit as five is to what? I've been calling them quints, but... --Ihope127 17:00, 27 August 2005 (UTC)
- Call it quits. -Splash 22:51, 28 August 2005 (UTC)
- Or a nybble and a bit. -Splash 23:07, 28 August 2005 (UTC)
- The proper name would be "quinary digits", so quints is probably as correct as anything. -Splash 23:07, 28 August 2005 (UTC)
- Or a nybble and a bit. -Splash 23:07, 28 August 2005 (UTC)
What is ten equal to? Tits?--Light current 22:57, 4 November 2005 (UTC)
Trivial bits
Though I've supported pop culture references in articles before, an arbitrary mention of the number of bits employed by the android from STTNG is far too much of a stretch here; any computer will have storage capacity measured in bits, and including one's particular favourite example just doesn't go. An article on "the bit" is extremely high-level by nature, because it's a fundamental concept of computation (and arguably, the universe). There's no natural place to include the Data mention (a "Trivia" heading doesn't count as natural), and that should have been a clue as to whether it was warranted in the first place. I'm afraid I'm going to have to revert the revert; please consider carefully before replacing the section. - toh 02:03, 4 November 2005 (UTC)
- agreed. — Omegatron 03:16, 4 November 2005 (UTC)
- I agree the memory capacity of Data (Star Trek) should not be in the Bit article, just as Data's height should not be in the meter article nor his mass in the gram article. -R. S. Shaw 07:33, 4 November 2005 (UTC)
- Hello! Define "natural" and arbitrary? Everything discussed therein is artificial, fictional or not. I included mention of Data's storage capacity since it was directly cited in the episode "The Measure of a Man" ("800 quadrillion bits") – when his status as a being vis-à-vis mere repository of information was challenged – and may be interesting, even trivial, for those who 'compare' fundamental concepts of technology with (fictional) "technology". Moreover, head on over to information, artificial intelligence, et al. for a description of concepts relating to said issues, and you'll see that its inclusion is completely relevant. This 'tid-bit' has also been classified as "trivia", so I do not see a reason to remove it.
- Many articles have other sections; if anything, as the topic of this article is such a "high-level" concept, this points to the deficiency of the current article and its need for expansion. Actually, perhaps such removals should be considered carefully (e.g., posing a request for comment here as was done after the fact) before doing so in the first place? Given the fact that this bit of info was around for quite some time before its removal, and my reasons above, I will be restoring this until compelling reasons are provided to remove it. E Pluribus Anthony 08:56, 4 November 2005 (UTC)
- I also agree that the information is of such overwhelming triviality to this article that it needn't belong. Put in an article about the character where it will have proper context. -Splash 12:54, 4 November 2005 (UTC)
- If you take a moment and actually glance at that article, it is in there, followed by a comparison. And what Wp policy excludes trivia from articles? To my knowledge: none. Actually, verifiability is the one I cite. Perhaps if various users would spend more time and effort on making substantive contributions and beefing up the present article instead of debating trivia and wikilinks that add value to it, it is (on the contrary) this debate which is elevating the topic beyond trivial. E Pluribus Anthony 13:01, 4 November 2005 (UTC)
- Excellent, it's already in the relevant article. That means we've no need at all to duplicate it here. Good, good. It can go then. -Splash 13:45, 4 November 2005 (UTC)
- If you take a moment and actually glance at that article, it is in there, followed by a comparison. And what Wp policy excludes trivia from articles? To my knowledge: none. Actually, verifiability is the one I cite. Perhaps if various users would spend more time and effort on making substantive contributions and beefing up the present article instead of debating trivia and wikilinks that add value to it, it is (on the contrary) this debate which is elevating the topic beyond trivial. E Pluribus Anthony 13:01, 4 November 2005 (UTC)
- Not good ... I don't think so. What's the rationale for exclusion here, based on Wp policy? Sorry! E Pluribus Anthony 13:48, 4 November 2005 (UTC)
- As others have already said, it is irrelevant to an article dealing with the abstract concept of a bit. The fictional storage capacity of one fictional computer is...well...of zero importance. Which, per the guidance in Misplaced Pages:Trivia#Trivia policy means it should get zero space in this article. Make List of storage capacity in bits of fictional computer characters or something. -Splash 14:01, 4 November 2005 (UTC)
- Not good ... I don't think so. What's the rationale for exclusion here, based on Wp policy? Sorry! E Pluribus Anthony 13:48, 4 November 2005 (UTC)
- Thanks for the policy; not only is it helpful, but it proves my point. It is a matter, then, of what is "interesting" and of mild "importance", for reasons cited above (e.g., computers and artificial intelligence): what's good for the goose isn't necessarily good for the gander. Your determination of nothingness, particularly regarding an abstract concept, is not mine; mine is at least verifiable, nor is its inclusion imbalanced which, if it were, would require edition or removal. In this respect, and in this article, this bit of information will remain until a compelling reason exists to remove it. Good day. E Pluribus Anthony 14:45, 4 November 2005 (UTC)
- In support of my assertion, I have added a real example for this 'abstract' concept ... moreso that present company. How do you like them ... bits? Enjoy! E Pluribus Anthony 15:32, 4 November 2005 (UTC)
- This is the wrong article. The information is already covered in the correct articles. — Omegatron 16:22, 4 November 2005 (UTC)
- This is precisely the correct article: it deals with information and units of storage in context. E Pluribus Anthony 17:10, 4 November 2005 (UTC)
- I'm inclined to agree with Omegatron, Splash, and Toh. In general I don't think 'trivia' sections really belong in encyclopedia articles, and it's too much of a stretch to include the examples that E Pluribus Anthony has presented. By all means, discuss the information capacity of the human brain in brain, and discussion Data's storage in Data. Those little disconnected factoids don't really belong here, in this general overview article. TenOfAllTrades(talk) 19:54, 4 November 2005 (UTC)
- Argh! This would not be so much of a grundging issue if it was addressed two months ago by various users, when the Data factoid was originally added. These bits of information, in this appropriate context, go to the very heart of what information is and how much of it readily apparent examples - us, and a popular sci-fi reference - exhibit. To do so now smacks of a 'fairweather', and slightly POV, attitude to this topic and the information presented thereof; in fact, the debate itself seems rather trivial and hypocritical, not the information in and of itself.
- Moreover, if users spent more time and effort enhancing articles, they likely wouldn't be confronted with the presentation of what are characterised as 'trivial' elements to spartan "top-level" articles to begin with, as even a prior user tried to 'spice' this one up a bit. E Pluribus Anthony 20:41, 4 November 2005 (UTC)
- So if people stopped adding irrelevant information to articles and revert warring it when it's removed and wasting everyone else's time in discussions about it, we could all spend more time adding quality information to relevant articles instead?
- I completely agree.
- Unfortunately, everyone doesn't think like that. Some people feel that their information is sooo important, they will "judiciously continue" adding it in spite of clear consensus to the contrary. — Omegatron 21:21, 4 November 2005 (UTC)
- I'm not exactly sure what you're arguing here. Removal of the trivia section can certainly be seen as 'enhancing' this article—the section has been criticized as unprofessional and disconnected. Writing isn't just about adding stuff; it's also about editing.
- That the section remained in the article for an extended period of time doesn't grant it some sort of special dispensation or tenure. Looking at the article history, between the time the 'Data' passage was added and when it was removed the article received next to no attention beyond the addition of interlanguage links and reversion of simple vandalism. The fact that new editors can take time to stumble across an article means that low-profile, uncontroversial articles will evolve very slowly.
- That Data (Star Trek) is a fond and familiar character for Wikipedians of a certain generation and scifi bent does not make him an exceedingly important fellow for a general-purpose encyclopedia article. The information is in Data's article where it belongs. A throwaway line created by some scriptwriter of a scifi TV series a decade or two ago doesn't really convey something useful about the concept of a 'bit'; it was just used as another chunk of technobabble filler. TenOfAllTrades(talk) 21:31, 4 November 2005 (UTC)
- I'm arguing that just because something is trivial (and this in accordance with cited policy, as being interesting and informative), does not obviate its inclusion here. I cite it as being potentially important to novice users who may not know a bit from a bite, or even a byte, and the examples put the definitions into context. I realise that its lengthy gestation on the page does not afford it special consideration but I think a true editor, given this forum, would find some way to incorporate it, morph it, or have it evolve within the article (as I also tried to do with an even more valid reference with the human brain), not to remove both outright; that seems questionable.
- (A segue: O, what's good for the good isn't good for the gander, so yes: it's relevant and sooo important given the desire to remove it; as well, I indicated I may include it if a clear consensus is not forthcoming, given opinions here. Next ...)
- While a beloved character, Data is unimportant per se; what is important, though, are the implications that this information regarding capacity has on one's understanding of bit, information, and AI to John Q. It's presence allows a user to put the concept into context, and its removal does not at all enhance understanding given the spartan nature of the article. I will field a public comment as suggested and be guided by the results of that. Thanks 10, for your continued, dispassionate counsel. Ta! E Pluribus Anthony 22:25, 4 November 2005 (UTC)
- I'm tempted to join in. But no I'll resist...I have other things to do!--Light current 22:41, 4 November 2005 (UTC)
- He he; likewise ... bring light unto the world! :) E Pluribus Anthony 22:45, 4 November 2005 (UTC)
- I agree; the trivia should not be included. But it seems like things have been resolved. Anyway, I added links to computer storage and transmission, to which these trivia are more relevant. --Mgreenbe 20:50, 15 November 2005 (UTC)
RfC on trivial bits
Hello! A Request for Comment has been submitted regarding the following:
Whether the following two items/section should be added to the article – whether they are irrelevant or necessary to provide conceptual context for (esp. novice) visitors.
Examples
- The information storage capacity of the human brain is somewhat subjective. In Carl Sagan's popular TV series and book Cosmos, the brain is indicated to contain some 10 neurons and each neuron has perhaps a thousand interconnecting dendrites. If a bit of information in the brain corresponds to one dendritic connection, Sagan posits the human brain may contain 100 terabits (10 bits), or 11 terabytes, of information.
- In the science fiction TV series Star Trek: The Next Generation, the android Data was built with an ultimate storage capacity of 800 quadrillion bits, or approximately 88 petabytes.
There has already been discussion in the #Trivial bits section above. Subsequent discussion should occur below, or elsewhere in this section if needed. Thanks! 11:49, 5 November 2005 (UTC)
- Generally I think trivia should be excluded from scientific and technology articles, unless it can add some useful knowledge. Then I suppose it ceases to be trivia. My opinion is that the stuff about the human brain could be included, but that about a ficticious android should be excluded. --Light current 13:42, 5 November 2005 (UTC)
- I generally concur for reasons indicated by Light current. Further to this, it may be prudent to include a superlative for contemporary computers and, if possible, the number of bits in the universe, etc. (since it may be a 'top-level' all-encompassing concept). As well, since the fictional reference is (albeit based on a little known 'throwaway' phrase) regarding a unique character in a popular TV series, it is nonetheless verifiable and provides a useful perspective about information pertaining to artificial intelligence (AI); thus, I believe it too should be included for comparison. E Pluribus Anthony 13:52, 5 November 2005 (UTC)
- This isn't an article about storage capacity. — Omegatron 15:21, 5 November 2005 (UTC)
- No: it's an article about quanta of information, requiring examples to illustrate the concept. E Pluribus Anthony 15:30, 5 November 2005 (UTC)
- You have asked for comments from third parties, so I dont think its useful to continue your argument here until you got some, is it? --Light current 15:37, 5 November 2005 (UTC)
- If they are germane to the topic, a response by anyone to comments made by anyone is perfectly justified; it would be an issue (and feel free to remove above comments) if they are inappropriate. That's it for me. :) E Pluribus Anthony 16:03, 5 November 2005 (UTC)
- You have asked for comments from third parties, so I dont think its useful to continue your argument here until you got some, is it? --Light current 15:37, 5 November 2005 (UTC)
- No: it's an article about quanta of information, requiring examples to illustrate the concept. E Pluribus Anthony 15:30, 5 November 2005 (UTC)
- As noted, I don't think the Data mention has a place here. The ostensible storage capacity of the human brain doesn't either, in part because the article isn't about storage capacity (as noted) and in part because any figure you could put there would be pretty much as fictional as the ST one, as in fact is the whole notion. The brain can't be particularly usefully described that way, at least not at this point in the history of psychology, unless it's in the same way that a toaster or a rock can be (ie. by viewing the world through the information theory lens, which is in fact something that could usefully be fleshed out a bit elsewhere in this article). I'd add that I'm dismayed that this particularly silly and obvious edit war has gone this far. Anthony, when you find this many people disagreeing with you on so many grounds (and there are several reasons the original note didn't fit the article which have nothing to do with its triviality), it's usually because you're mistaken, not because you're a lone visionary. I speak from experience. ;) - toh 08:58, 6 November 2005 (UTC)
- We agree to disagree. Moreover, the last time I checked, the examples have not been reverted recently, so the edit 'war' is actually a detente pending this RfC. Lastly: while I will gladly admit potential error (as others should), viewing topics through a solitary lens can (to extend your visual metaphors) be construed as being myopic: another user (somewhat removed from prior discussion) has at least found value in the brain example, so the discussion isn't necessarily a fruitless one. I can deal with that ... and have very thick glasses to prove it. :) Take care! E Pluribus Anthony 10:16, 6 November 2005 (UTC)
- Hi all, I am just dropping in from the RfC page, so I am not sure I understand the nature of the dispute, but I do have a comment about one of the two examples suggested above. I suggest that you avoid mentioning the information storage capacity of the human brain, since this raises all kinds of presumably extraneous issues, such as
- precisely how mammalian brains store/process information remains mysterious,
- there are various possible definitions of information, and Shannon's may not always be the most appropriate one, particularly in biology. (For instance, biologists who ought to know better often refer without comment to "the information content of an E. Coli", apparently assuming this means something defined entirely in terms of the genome, without considering that the genome of an E. Coli cannot a bacteria make without some infrastructure such as cell membranes, raw materials, and the like, which suggests that something essential might be missing from any definition which only attempts to take account of the genome itself.)
- You can probably find another example which makes your point without raising thorny issues like these.---CH (talk) 01:37, 7 November 2005 (UTC)
- Thanks for the feedback; I think your concern is partially covered off with upfront notation/qualification of the factoid being "subjective" (perhaps the wording can be massaged). Further to that, the examples provided are ones that a general user can readily understand and can be verified; however, I'd invite, be willing, and able to incorporate other examples (popularly cited) and or discussions regarding notions that quantify information/bits in toto of any biological or other system (e.g., the universe?). It is partially for these reasons that such examples are valid and should be included. Besides: what rose (and bits thereof) doesn't have thorns? :) That's my two ... bits worth. E Pluribus Anthony 03:04, 7 November 2005 (UTC)
- Hi all, I am just dropping in from the RfC page, so I am not sure I understand the nature of the dispute, but I do have a comment about one of the two examples suggested above. I suggest that you avoid mentioning the information storage capacity of the human brain, since this raises all kinds of presumably extraneous issues, such as
Re the RfC: Anthony, this article is not actually about "quanta of information". That article is the data article (see also Wiktionary:data) (yes, I appreciate the irony :d). How much data the human brain can hold can be estimated in terms of bits, but that should be on the memory or human brain articles. Consider:
- "...The base is partly a matter of convention, like using centimetres instead of inches, partly a matter of convenience, since it is useful when dealing with digital devices that use binary codes to represent data." - 2.1 The mathematical theory of communication, Semantic Conceptions of Information (You should read the paragraphs that come before the quote for context)
Also check out What is Mathematics? which I have a feeling you will probably find very interesting.
As for any discussion about whether the universe is basically a big (digital) computer, that should go on Digitalism (or Digital philosophy, as both articles are basically stubs and maybe they should be merged). You might also want to take a look at information science, information theory, mathematical physics, biological neural networks, cognitivism (psychology), and quantum mind just to get an idea of "what's out there". These are pretty specific articles, so you should try to get a fuller perspective by browsing the categories they are in too.--Ben 12:57, 12 November 2005 (UTC)
- Hi there! Thanks for the helpful information, far more so than the comments (methink) of prior detractors. While I believe that this should still be noted (but not necessarily expanded per se) in the current article, with detail in articles you cite et al., the majority (but not totality) of commenting users believe this information would be more appropriately placed elsewhere. I think it prudent to allow more commentary, and time, from others before unequivocally doing so. Thanks again! E Pluribus Anthony 15:04, 12 November 2005 (UTC)
A bit is a binary digit and I judge it to be essentially a mathematical concept for representing place values of a binary number - in the same way as decimal digits are a concept for representing the place values of a decimal number. So I do not think it is a storage concept, and agree with Ben's well put suggestion's on suitable articles for information about memory. Fuzzyslob 07:48, 13 November 2005 (UTC)
Agree that the trivia are irrelevant; added a link to computer storage and transmission, the former of which may be an appropriate place. --Mgreenbe 21:09, 15 November 2005 (UTC)
RfC conclusion
Thank you all for your comments. I believe a consensus is now apparent that the 'trivial bits' at the centre of this issue would be more appropriate elsewhere. To that end, I'll close this RfC ... and will shortly find these puppies a real home. :) E Pluribus Anthony 13:31, 22 November 2005 (UTC)
64-bit console
So a 64-console(or 128, 256 etc. etc.) is a console with a 64-bit byte?
- "N-bit" in game consoles (and for that matter general-purpose computers) is often more of a marketing term than a technical one, but you're right that the article needs to talk about that popular usage, probably in a new section. Most of the details of what it's supposed to mean are described at central processing unit andprocessor register and word. Some of it is also covered at byte, which is a value whose width has varied, but never as large as 64 bits. The width of a processor can be thought of as the largest number that processor can work on in one operation, analogous to the number of fingers and toes you happen to have to count on. In some cases processor width contributes to performance efficiency, hence the marketing usage. - toh 19:34, 11 November 2005 (UTC)
- Things like 16-bit have articles, by the way. — Omegatron 00:02, 12 November 2005 (UTC)
kilo
Shouldn't kilobit be written as kb (not Kb) and kilobyte - kB (not KB) because the symbol of kilo- is 'k', not 'K': Kilogram - kg etc...???
- Yes, using 'K' instead of 'k' for 'kilo' is sloppy. Much sloppier, though, is the use of kilo for 1024 instead of 1000. See binary prefix. -R. S. Shaw 19:52, 15 April 2006 (UTC)
The ambiguity has been solved in 1998 by the IEEE: see KiB. 151.46.147.128 13:39, 17 January 2007 (UTC)
Origin of the word "Bit"
Isn't it more likely that the term comes from "Pieces of Eight"? Each "piece" was called a bit. A quarter was two bits. It seems much more likely to me that this is the source of the computer term than the rather strained (IMO) abbreviation (or "portmandeau word") from Binary digIT. Dduggan47 13:49, 4 April 2006 (UTC)
- Is that you don't believe Claude Shannon, a very early user of the term, that he got it from John Tukey? Or is it that you think Tukey, a statistician who is credited with creating a fair number of terms, didn't make it up? Or that Tukey was just humming "Shave and a haircut, two bits" and decided to invent a short name for a binary digit and thought "I know, I'll call it a 'shave' – No, maybe 'bit' would be better"? You find hard to believe that BIT could be a contraction of BInary digiT? IMHO, that's a lot more plausible than pieces of eight, awwk, pieces of eight. -R. S. Shaw
- It isn't a matter of not believing in the sense of thinking the guy lied. I remember an occasion on which I made a comment to a friend that was very funny and, I thought, very original. I later found that my comment was not original at all and was patterned on something I'd undoubtedly read before but had forgotten about, at lease consciously. In this case, bits are an eighth part of a Spanish coin. The coin, being gold, could be cut with a knife. You could cut it in half, you could cut the half in half and get a quarter, and you could cut the quarter in half and get what they called a bit. AIUI a bit was small enough that it could not easily be cut in half again. So, we've got something called a bit which is an eighth of a whole unit and which cannot be further subdivided. The word came into common usage as a term for an eighth of a dollar. Then, at a later date, we have a fellow who is said to have invented the word from BInary digiT and given it the meaning of an indivisible one eighth part of a whole unit. I guess I'd put the question back to you as to what's the more plausible origin of the word. I understand your argument, I just don't find that story likely. It seems clear to me that the word "bit" meaning an indivisible one eighth of something long predates Tukey and computers. Dduggan47 11:36, 14 April 2006 (UTC)
- You seem fundamentally confused about the meaning of 'bit' as it relates to the binary number system (the subject of the article). It has nothing to do with being "an eighth of a whole unit". One might as well claim a bit is a quarter of a whole unit or a thirteenth of a whole unit. A bit is not a fraction of any particular unit more than it is of any other unit. A bit is a binary digit, that is, a item of information that can take on exactly one of two values. This meaning of 'bit' is tightly coupled to the value 2, but has no standout relationship to the value 8 (certainly no more than 4, or 16, 32, 64, ...).
- Also, Tukey did not give 'bit' the "meaning of an indivisible part of a whole unit". 'Bit' doesn't mean that any more than 'decimal digit' means "indivisible part of a whole unit". The meaning Tukey assigned to the existing word 'bit' was, please note, "binary digit".
- The word 'bit' had many meanings before it was first used to mean 'binary digit'. Yes, it was was used in the 1600's and later to mean a fraction of a dollar, variously an eighth, tenth, or sixteenth (Oxford English Dictionary). But the word was used centuries before that to mean things like 'morsel' or 'small piece', and means many other things (OED). It's clearly a nice coincidence that the contraction 'bit' has overtones of "small thing", but that's less important than it's being a single syllable and being mnemonically related to "binary digit". -R. S. Shaw 19:30, 15 April 2006 (UTC)
- Thank you for your reply.
- You are correct that despite the first sentence of the article, I persisted in thinking of the word in a computer sense rather than in a mathematical one. I've never encountered bit as a synonym for binary digit before. More often in my experience it's more like The Amercan Heritage Dictionary's "A fundamental unit of information having just two possible values, as either of the binary digits 0 or 1.". It's in the sense of being "a fundamental unit of information" that it's indivisible. So ... yes, as you point out, I was "fundamentally" confused about the usage here. (I don't have the Oxford Dictionary but the usage of bit as any specific fraction other than 1/8 is a new one on me too ... and I've looked. I'm not doubting you, I've just never seen it before.)
- You've answered my original question and brought me most of the way to your view (which is the accepted etymology). I'll let it drop with just tone more word. I think a site called the Online Etymological Dictionary may have the best take on it (http://www.etymonline.com/index.php?term=bit). The non-mathemtatical definition ("small piece", money, etc.) is given and then definition 2 says it's a contraction of "binary digit" coined by Tukey "probably chosen for its identity with ". In other words, Tukey did coin the term by contracting "bnary digit", but he also probably knew the history of the word he was coining. 70.20.24.129 21:46, 15 April 2006 (UTC)
- Well, etymonline's listing for "byte" is wrong, so I wouldn't trust them too far. They say "1964, Amer.Eng., see bit" but it was coined in July 1956 by Dr. Werner Buchholz. And the "pieces of eight" is purely speculation, and wrong. Bits have not always been 1/8 of a byte. The original meaning of "byte" was any small discrete unit made of a number of bits. Some systems had 6-bit bytes, for instance. Also, considering the word "bit" was coined in 1948 and "byte" was coined in 1956, byte is clearly an extension of "bit"; not the other way around. — Omegatron 14:12, 24 March 2007 (UTC)
Information theory bits
Bits are mentioned with reference to information theory as a measure of information entropy, but are not defined for this use in the article. There is a difference between the use of "bit" in noting a 0 or 1, and the use of "bit" to measure information. If a file with 1000 0s and 1s losslessly compressed to a file of 500 0s and 1s, then the original file, although having 1000 bits in the 0/1 sense, has at most 500 bits of information entropy, since information is not destroyed by lossless compression, and a file can have no more information theoretical bits than it can 0/1 bits. I'm not sure whether there needs to be a bit (information theory) article or another section here, but, as it is, the definition of bit here is incomplete. Calbaer 04:00, 5 June 2006 (UTC)
- I see this is a very old comment, but it's still relevant nonetheless. The article does not make it clear that the word "bit" has more than 1 meaning. The article gives the 2 definitions, but doesn't make it clear that those are different things.
- A bit is a representation of a system with binary output, and it is also a unit. These are different meanings, it is not clear in the article, leading to questions about if fractional number of bits are possible. A new section expliciting the difference between bit as an abstraction of a binary system, and bit as used in theory of information is of utmost importance.
- 177.68.225.247 (talk) 22:28, 24 March 2017 (UTC)
I too think this comment is quite relevant and important. The double meaning of bits is quite unfortunate. Bits as binary digits are concerned with data representation, and bits as information theory units are concerned with information/surprisal/entropy. These are very different concepts that are heavily intermixed in this article. It might be helpful to mention the alternate term for information theory bits: Shannons. See the article Shannon_(unit) for a little more detail. Tyler Streeter (talk) 16:17, 3 July 2019 (UTC)
- Shannon said "Bits as binary digits" are "bits as information theory units". --Masataka Ohta (talk) 07:49, 13 April 2023 (UTC)
I will add an etymology section to explain two meanings of a bit and a binary digit referring the original paper of Shannon. Section 9 of the paper is on source coding theorem, where Shannon himself uses a "binary digit" not as a unit of information but as a symbol. --Masataka Ohta (talk) 07:49, 13 April 2023 (UTC)
Representation
It's already been tagged as "citation needed", but as an (admittedly poor) electrical engineer, I must say that this section is misleading at best and possibly flat-out wrong. I suppose what the author was trying to get at was the difference between Emitter Coupled Logic (ECL) and CMOS or TTL, but the statement made about 0V rarely being used to represent a logical 0 is just plain false. Similarly, implying that "slightly older devices" (whatever that entirely unspecific statement means) are no longer used is incorrect as well. Devices like those in the 7400 series are CMOS devices that definitely use a 0V-5V scheme and are certainly still widely used today. 65.96.178.66 06:15, 18 December 2006 (UTC)mmoskwa
Bit and Byte
I added that information because after skimming the article on the bit, I didn't see any OBVIOUS mention on the byte being 8 bits in length as is used today in modern computers. Read the article on the byte if confused. It also states towards the middle, in so many words, on the ubiquitous nature of the byte being 8 bits in length (look up the word "ubiquitous" in Webster's dictionary). Misplaced Pages redirects ubiquitous to omnipresence, but still talks about "ubiquity" in the beginning of the article. Also see bisnotB.com for an obvious entry on today's computer usage of the bit versus byte. It should also be mentioned here that a bit is a lowerecase "b" and maybe here or in the byte section, a byte is a capital B. As is used in downloads and transfers, most of the time, the "KB/s" or "kB/s" is used (Kilobytes per second): example... 598KB/s ... which happens to be my max steady download rate. When talking about mp3s, "Kb" or "kb" is used: 128kbs (128kb/s) or 160kbs (160kb/s). If you don't believe me on this, right click an MP3 file on your computer if you have one, and check out it's properties in the menu, then click on 'summary', and then 'advanced' if in the 'simple' view. —The preceding unsigned comment was added by Bill Riojas Mclemore (talk • contribs) 14:49, 14 April 2007 (UTC).
hello —Preceding unsigned comment added by 203.125.157.117 (talk) 07:26, 7 September 2007 (UTC)
In the international standard IEC (IEC) 60027-2 2005, the following symbols are recommended for use in the electrical and electronic fields:
'Bit' to indicate a bit; — Preceding unsigned comment added by 77.38.193.104 (talk) 21:39, 8 October 2020 (UTC)
Millibit
Is millibit/microbit/nanobit/picobit for real?Anwar (talk) 19:05, 10 May 2008 (UTC)
- If you are receiving data from your deep-space probe via a weak radio signal, and it takes 2 min to receive 60 bits, your data transfer rate is 500 millibits per second. Indefatigable (talk) 20:07, 12 May 2008 (UTC)
- LOL! So it is abstract. But that's a very mischievous way of counting. For instance, consider:
statement 1 - the book is on the table
statement 2 - the table is on the floor
conclusion - the book is on the floor.Anwar (talk) 13:38, 22 May 2008 (UTC)
- LOL! So it is abstract. But that's a very mischievous way of counting. For instance, consider:
Deleted "cute" (non)definition
I deleted this sentence, as it seems to contain zero bits of useful information:
Gregory Bateson defined a bit as "a difference which makes a difference". ([http://plato.acadiau.ca/courses/educ/reid/papers/PME25-WS4/SEM.html Social Systems) |
All the best, --Jorge Stolfi (talk) 06:46, 2 July 2009 (UTC)
Introduction
The first sentence reads:
- A bit (a contraction of binary digit) is the basic unit of information in computing and telecommunications; it is the amount of information stored by a digital device or other physical system that exists in one of two possible distinct states.
A unit of information isn't an amount. You might say it is a capacity, but even then ...
The sentence tries to define it twice. Once correctly as an abstract, secondly "incorrectly?" as the physical manifestation. HonestIntelligence (talk) 15:33, 2 January 2012 (UTC)
- Both statements are true and essential to understanding what a "bit" is. Perhaps a unit cannot be an amount, but a quantity of one of a unit can be described as equivalent to an amount described by some other means. "A cup is a unit of measure in the English system of units, primarily used in food preparation. It is the amount of liquid stored in a container of about 236 millileters volume."
- I think it's fine as it is. Perhaps you can suggest an alternate wording (but again, both statements are essential to understanding, so please do not simply delete either). Jeh (talk) 20:18, 2 January 2012 (UTC)
Hmm
My take is: A bit is something which represents either 1 or 0 (one or zero). The "something" may be implemented in a variety of systems from computing to telecommunications to DNA.
It certainly isn't "an amount", as said previously, it is a capacity.
The lead seems overweighty, perhaps a lot can be moved into "Representation"? It "can" be... seems a recipe to infinity!
I will have a think! :) HonestIntelligence (talk) 11:50, 28 May 2012 (UTC)
OK. Done some editing. HonestIntelligence (talk) 12:21, 28 May 2012 (UTC)
Pieces of eight
I was just reading a fascinating book called "Machine Language for Beginners", which deals with computers that use 6502 processors, and it gives a fascinating explanation that the definition of bit is NOT binary digit but derives from pieces of eight! Anyone able to verify this? -121.216.109.33 (talk) 11:17, 17 April 2012 (UTC)
- I am pretty sure that is a false etymology, as the word bit for a binary digit was used long before the firm establishment of a byte being 8 bits. Indefatigable (talk) 21:04, 18 April 2012 (UTC)
Representation
I moved a paragraph from the lead here, however the subheading of Transmission and processing says little about either. I'll do something "place holder-y" for the moment! HonestIntelligence (talk) 12:27, 28 May 2012 (UTC)
Bits are indivisible
There seems to be some confusion both here and on the Byte wiki page about the bit being divisible and what the names are for that. However, bits are not divisible except when being used as a rate such as "1/2 bits per second" that can otherwise be written as "1 bit per 2 seconds" or "2 seconds per bit". This information is missing from the Bit page. Of course any experienced engineer would know this already, but not many engineers come looking on a wiki page for the names of fractions of bits. This must be written out here and should be written out in the Byte page. This is Madness 05:44, 30 June 2012 (UTC)
I agree. As you mentioned, and in general information theory and mathematics, a bit is just as divisible as any other numeric value. But in practice, a bit is always treated as an atomic unit. So any mention of divisibility or fractions is silly and should be removed from any computing or engineering article in order to prevent confusion. The closest thing to an actual fractional value would be when considering the amount of energy required to pass through an electronic gate in order to register as a bit, and even then it becomes a matter of energy versus data measurement. 174.20.107.90 (talk) 17:39, 4 September 2015 (UTC)
- When referring to a representation of a system that can have 1 of 2 possible states, yes a bit is indivisible. But when used as a unit of information (the shannon), you may obtain non-integer values of bits (for example the outcome of a fair, six-sided dice has ~2.6 bits of information (shannons)).
- The problem is that the word "bit" has 2 different meanings, 2 different concepts. This kind of confusion is the reason I believe a distinction between "bit as a binary outcome" and "bit as an unit of information" (also known as shannon) needs to be clearly addressed in the article.
- 177.68.225.247 (talk) 22:11, 24 March 2017 (UTC)
Values
Any Reference Available for
Saganbyte = 1024 Geobyte Pijabyte = 1024 Saganbyte Alphabyte = 1024 Pijabyte Kryatbyte = 1024 Alphabyte Amosbyte = 1024 Kryatbyte Pectrolbyte = 1024 Amosbyte Bolgerbyte = 1024 Pectrolbyte Sambobyte = 1024 Bolgerbyte Quesabyte = 1024 Sambobyte Kinsabyte = 1024 Quesabyte Rutherbyte = 1024 Kinsabyte Dubnibyte = 1024 Rutherbyte Seaborgbyte = 1024 Dubnibyte Bohrbyte = 1024 Seaborgbyte Hassiubyte = 1024 Bohrbyte Meitnerbyte = 1024 Hassiubyte Darmstadbyte = 1024 Meitnerbyte Roentbyte = 1024 Darmstadbyte Coperbyte = 1024 Roentbyte...
?
if yes then please let me know
Sopan Patil (talk) 12:48, 8 January 2013 (UTC)
- The oldest reference I found for the list from Saganbyte to CoperByte is: "Several Data Bytes Unit Conversions". May 2000. Retrieved November 23, 2018. Wells50 (talk) 20:21, 23 November 2018 (UTC)
Additional Prefix Info
This is an excellent question. I've seen these values floating all over the web on various blog posts, all dated from around 2013, with no attribution. I wonder if some random guy just made the list on a lark and it somehow spread around, due to lack of official prefixes. The random jumps between names of minerals and scientific figures, and things like "Infinity" and "Website" seems suspect. A list of corresponding shorthand (like GB for gigabyte) would be nice as well (presumably Brontobyte is BB but Geobyte can't be GB). Here's a bit more info I've found:
Brontobyte and Geobyte (aka Geopbyte / Gegobyte) - Source: Shantanu Gupta, director of Connected Intelligent Solutions for Intel's Data Center and Connected Systems Group. Proposed standard as of 2012, discussed since 1991. De-facto, informal standard.
Saganbyte and Pijabyte - Seem to be accepted by some, but unknown origin and limited real world use.
Alphabyte through Websitebyte - Unknown origin, and no real world use outside of appearing in lists.
Here's the full list of Byte prefixes floating around the web, in ascending order (43 total). Please note that I'm not promoting these, just relaying them.
(0) bit, (1) Byte, Kilobyte, Megabyte, Gigabyte, Terabyte, Petabyte, Exabyte, Zettabyte, Yottabyte, (10) Brontobyte, Geobyte, Saganbyte, Pijabyte, Alphabyte, Kryatbyte, Amosbyte, Pectrolbyte, Bolgerbyte, Sambobyte, (20) Quesabyte, Kinsabyte, Rutherbyte, Dubnibyte, Seaborgbyte, Bohrbyte, Hassiubyte, Meitnerbyte, Darmstadbyte, Roentbyte, (30) Coperbyte, Koentekbyte, Silvanikbyte, Golvanikbyte, Platvanikbyte, Einstanikbyte, Emeranikbyte, Rubanikbyte, Diamonikbyte, Amazonikbyte, (40) Nilevanikbyte, Infinitybyte, Websitebyte.
Sources: (web search reveals many more)
97.116.45.199 (talk) 19:42, 25 November 2014 (UTC)
References
- http://itknowledgeexchange.techtarget.com/storage-disaster-recovery/exabyte-zettabyte-yottabyte-then-what-opinions-vary/
- http://www.technolamp.co.in/2010/12/computer-bit-size-chart-from-bit-to.html
- https://answers.yahoo.com/question/index?qid=20090219174202AAbjkY4
- http://passion4profession.blogspot.com/2013/01/1024-coperbyte.html
- http://sanjaal.com/java/tag/hassiubyte/
- https://www.ddicomputer.com/?r=support/ViewArticle&id=397
- http://thehackeruniversity.blogspot.com/2013/01/file-sizes.html
- http://metalop.com/2013/10/16/binary-conversions-you-wont-know-about/
- http://hirenchavdahdk.blogspot.com/2013/10/hard-disk-storage-space-units.html
- http://www.codeloopers.com/storage-and-usb-devices-f22/what-are-smallest-and-largest-data-bytes-or-byte-scales-t12.html
- When SI or IEC or even JEDEC recognizes these, sure. Until then, no. This isn't the Urban Dictionary, nor Merriam-Webster (i.e. we're not here to document this year's new fad words). Jeh (talk) 23:03, 4 September 2015 (UTC)
2-Bit
Don't we have the 2-bit computing environment?
We have long list of 2-bit logic subject for computing Boolean algebra here. I haven't seen 2-bit computing out there yet. — Preceding unsigned comment added by Rjluna2 (talk • contribs) 12:53, 23 July 2013 (UTC)
Older origin for the word "bit"
This MOOC (at 6"55) relates the origin of "bit" to Leo Szilard's essay about information theory and Maxwell's Demon in 1929, in which he wrote: "A bit of information is the amount of information needed to answer a fast/slow question, or any yes/no question". Another source for this can be found on the Leo Szilard Online page.
- The following paper by him on 1929 "ON THE DECREASE OF ENTROPY IN A THERMODYNAMIC SYSTEM BY THE INTERVENTION OF INTELLIGENT BEINGS" mentions Maxwell's demon but not "bit" at all. As the paper was originally written in German, where "digit" is "Ziffer", there shouldn't be any "bit". Moreover, the paper has, without specifying the base, many expressions of to represent amount of information, which means the base is unlikely to be 2, which, in turn, means the unit of information in the paper is not a bit. --Masataka Ohta (talk) 07:32, 13 April 2023 (UTC)
Shouldn't we add this etymology as the link between the old meaning and the computer term ? --Martin Prunières (talk) 06:52, 24 November 2016 (UTC)
- Who is "we"? Be WP:BOLD Jeh (talk) 07:02, 24 November 2016 (UTC)
- What do you think? 77.221.89.89 (talk) 18:14, 9 January 2023 (UTC)
Is it a common misconception that instructions are ones and zeroes?
In order to meet the criteria at Talk:List of common misconceptions, the article Bit would need to say that it is a common and current misconception that people today think that electronic computers are storing and operating on ones and zeroes, as numerals, when in fact these are high and low voltages which can be thought of a representations of true and false or yes and no or many other things depending on the logical operation they are doing. Conceptualizing them as ones and zeroes happens at a higher level. This is commonly taught introductory curriculum as a step towards a more subtle understanding, as the hydraulic analogy is an oversimplification of how electric circuits work.
So can we find sources to support this, and add it to this article? --Dennis Bratland (talk) 20:57, 18 December 2017 (UTC)
- Nope. As pointed out at the other talk page, that's an abstraction, not a misconception. Also, it's the best possible abstraction, because it helps explain how bitwise operations work. If we used Left/Right to indicate the two states of a switch, it would not be easy to explain how something like XOR works. ᛗᛁᛟᛚᚾᛁᚱPants Tell me all about it. 23:11, 18 December 2017 (UTC)
- With positive logic, higher voltage is 1, whereas, with negative logic, higher voltage is 0, as is explained in Logic level. It is a low level convention. But such convention is often meaningless, for example, with Emitter-coupled logic. --Masataka Ohta (talk) 05:57, 12 April 2023 (UTC)
Recent addition
@Masataka Ohta, I reverted your addition to the article. The lede is supposed to be a short and very readable summary of the whole article; see MOS:LEDE. The content you added, which I found rather difficult to follow, belongs, if anywhere, somewhere in the article body – possibly in the history section? You should also ensure that everything you write is supported by references to reliable sources and is not based on your own supposition or original research. Best, Wham2001 (talk) 09:30, 10 April 2023 (UTC)
I will add a new section for etymology referencing Shannon's original paper to have defined "bit" with two meanings (as a unit and as a symbol) and clarifies such misconceptions as a bit, as a unit, could have a state represented by 0/1, even though it's bit as a storage/symbol which has state to store/encode bit values. --Masataka Ohta (talk) 05:43, 12 April 2023 (UTC)
Revisions to lead
This new addtion to the lead is confusing and poorly explained, and it seems like it relies on a paper written by the editor making the change - (see WP:COI, this is a problem in itself). If you are trying to mention the difference between a bit as used in digital communication and a Shannon (unit), so it would be redundant as well. MrOllie (talk) 12:59, 8 November 2024 (UTC)
Categories: