1
0
mirror of https://github.com/golang/go synced 2024-11-25 07:57:56 -07:00

pkg: spelling tweaks, A-H

R=ality, bradfitz, rsc, dsymonds, adg, qyzhai, dchest
CC=golang-dev
https://golang.org/cl/4536063
This commit is contained in:
Robert Hencke 2011-05-18 13:14:56 -04:00 committed by Russ Cox
parent b256358008
commit c8727c81bb
57 changed files with 120 additions and 120 deletions

View File

@ -164,9 +164,9 @@ func (oi ObjectIdentifier) Equal(other ObjectIdentifier) bool {
return true return true
} }
// parseObjectIdentifier parses an OBJECT IDENTIFER from the given bytes and // parseObjectIdentifier parses an OBJECT IDENTIFIER from the given bytes and
// returns it. An object identifer is a sequence of variable length integers // returns it. An object identifier is a sequence of variable length integers
// that are assigned in a hierarachy. // that are assigned in a hierarchy.
func parseObjectIdentifier(bytes []byte) (s []int, err os.Error) { func parseObjectIdentifier(bytes []byte) (s []int, err os.Error) {
if len(bytes) == 0 { if len(bytes) == 0 {
err = SyntaxError{"zero length OBJECT IDENTIFIER"} err = SyntaxError{"zero length OBJECT IDENTIFIER"}
@ -269,7 +269,7 @@ func isPrintable(b byte) bool {
b == ':' || b == ':' ||
b == '=' || b == '=' ||
b == '?' || b == '?' ||
// This is techincally not allowed in a PrintableString. // This is technically not allowed in a PrintableString.
// However, x509 certificates with wildcard strings don't // However, x509 certificates with wildcard strings don't
// always use the correct string type so we permit it. // always use the correct string type so we permit it.
b == '*' b == '*'

View File

@ -10,7 +10,7 @@ import (
"strings" "strings"
) )
// ASN.1 objects have metadata preceeding them: // ASN.1 objects have metadata preceding them:
// the tag: the type of the object // the tag: the type of the object
// a flag denoting if this object is compound or not // a flag denoting if this object is compound or not
// the class type: the namespace of the tag // the class type: the namespace of the tag

View File

@ -351,7 +351,7 @@ func marshalBody(out *forkableWriter, value reflect.Value, params fieldParameter
startingField := 0 startingField := 0
// If the first element of the structure is a non-empty // If the first element of the structure is a non-empty
// RawContents, then we don't bother serialising the rest. // RawContents, then we don't bother serializing the rest.
if t.NumField() > 0 && t.Field(0).Type == rawContentsType { if t.NumField() > 0 && t.Field(0).Type == rawContentsType {
s := v.Field(0) s := v.Field(0)
if s.Len() > 0 { if s.Len() > 0 {
@ -361,7 +361,7 @@ func marshalBody(out *forkableWriter, value reflect.Value, params fieldParameter
} }
/* The RawContents will contain the tag and /* The RawContents will contain the tag and
* length fields but we'll also be writing * length fields but we'll also be writing
* those outselves, so we strip them out of * those ourselves, so we strip them out of
* bytes */ * bytes */
_, err = out.Write(stripTagAndLength(bytes)) _, err = out.Write(stripTagAndLength(bytes))
return return

View File

@ -736,7 +736,7 @@ var deBruijn64Lookup = []byte{
func trailingZeroBits(x Word) int { func trailingZeroBits(x Word) int {
// x & -x leaves only the right-most bit set in the word. Let k be the // x & -x leaves only the right-most bit set in the word. Let k be the
// index of that bit. Since only a single bit is set, the value is two // index of that bit. Since only a single bit is set, the value is two
// to the power of k. Multipling by a power of two is equivalent to // to the power of k. Multiplying by a power of two is equivalent to
// left shifting, in this case by k bits. The de Bruijn constant is // left shifting, in this case by k bits. The de Bruijn constant is
// such that all six bit, consecutive substrings are distinct. // such that all six bit, consecutive substrings are distinct.
// Therefore, if we have a left shifted version of this constant we can // Therefore, if we have a left shifted version of this constant we can

View File

@ -84,7 +84,7 @@ func (z *Rat) Num() *Int {
} }
// Demom returns the denominator of z; it is always > 0. // Denom returns the denominator of z; it is always > 0.
// The result is a reference to z's denominator; it // The result is a reference to z's denominator; it
// may change if a new value is assigned to z. // may change if a new value is assigned to z.
func (z *Rat) Denom() *Int { func (z *Rat) Denom() *Int {

View File

@ -284,7 +284,7 @@ func (bz2 *reader) readBlock() (err os.Error) {
repeat := 0 repeat := 0
repeat_power := 0 repeat_power := 0
// The `C' array (used by the inverse BWT) needs to be zero initialised. // The `C' array (used by the inverse BWT) needs to be zero initialized.
for i := range bz2.c { for i := range bz2.c {
bz2.c[i] = 0 bz2.c[i] = 0
} }
@ -330,7 +330,7 @@ func (bz2 *reader) readBlock() (err os.Error) {
if int(v) == numSymbols-1 { if int(v) == numSymbols-1 {
// This is the EOF symbol. Because it's always at the // This is the EOF symbol. Because it's always at the
// end of the move-to-front list, and nevers gets moved // end of the move-to-front list, and never gets moved
// to the front, it has this unique value. // to the front, it has this unique value.
break break
} }

View File

@ -68,7 +68,7 @@ func newHuffmanTree(lengths []uint8) (huffmanTree, os.Error) {
// each symbol (consider reflecting a tree down the middle, for // each symbol (consider reflecting a tree down the middle, for
// example). Since the code length assignments determine the // example). Since the code length assignments determine the
// efficiency of the tree, each of these trees is equally good. In // efficiency of the tree, each of these trees is equally good. In
// order to minimise the amount of information needed to build a tree // order to minimize the amount of information needed to build a tree
// bzip2 uses a canonical tree so that it can be reconstructed given // bzip2 uses a canonical tree so that it can be reconstructed given
// only the code length assignments. // only the code length assignments.

View File

@ -185,7 +185,7 @@ func (w *huffmanBitWriter) writeBytes(bytes []byte) {
_, w.err = w.w.Write(bytes) _, w.err = w.w.Write(bytes)
} }
// RFC 1951 3.2.7 specifies a special run-length encoding for specifiying // RFC 1951 3.2.7 specifies a special run-length encoding for specifying
// the literal and offset lengths arrays (which are concatenated into a single // the literal and offset lengths arrays (which are concatenated into a single
// array). This method generates that run-length encoding. // array). This method generates that run-length encoding.
// //
@ -279,7 +279,7 @@ func (w *huffmanBitWriter) writeCode(code *huffmanEncoder, literal uint32) {
// //
// numLiterals The number of literals specified in codegen // numLiterals The number of literals specified in codegen
// numOffsets The number of offsets specified in codegen // numOffsets The number of offsets specified in codegen
// numCodegens Tne number of codegens used in codegen // numCodegens The number of codegens used in codegen
func (w *huffmanBitWriter) writeDynamicHeader(numLiterals int, numOffsets int, numCodegens int, isEof bool) { func (w *huffmanBitWriter) writeDynamicHeader(numLiterals int, numOffsets int, numCodegens int, isEof bool) {
if w.err != nil { if w.err != nil {
return return

View File

@ -11,7 +11,7 @@ import (
) )
// pipe creates two ends of a pipe that gzip and gunzip, and runs dfunc at the // pipe creates two ends of a pipe that gzip and gunzip, and runs dfunc at the
// writer end and ifunc at the reader end. // writer end and cfunc at the reader end.
func pipe(t *testing.T, dfunc func(*Compressor), cfunc func(*Decompressor)) { func pipe(t *testing.T, dfunc func(*Compressor), cfunc func(*Decompressor)) {
piper, pipew := io.Pipe() piper, pipew := io.Pipe()
defer piper.Close() defer piper.Close()

View File

@ -22,7 +22,7 @@ type Interface interface {
} }
// A heaper must be initialized before any of the heap operations // A heap must be initialized before any of the heap operations
// can be used. Init is idempotent with respect to the heap invariants // can be used. Init is idempotent with respect to the heap invariants
// and may be called whenever the heap invariants may have been invalidated. // and may be called whenever the heap invariants may have been invalidated.
// Its complexity is O(n) where n = h.Len(). // Its complexity is O(n) where n = h.Len().

View File

@ -284,7 +284,7 @@ func (curve *Curve) Marshal(x, y *big.Int) []byte {
return ret return ret
} }
// Unmarshal converts a point, serialised by Marshal, into an x, y pair. On // Unmarshal converts a point, serialized by Marshal, into an x, y pair. On
// error, x = nil. // error, x = nil.
func (curve *Curve) Unmarshal(data []byte) (x, y *big.Int) { func (curve *Curve) Unmarshal(data []byte) (x, y *big.Int) {
byteLen := (curve.BitSize + 7) >> 3 byteLen := (curve.BitSize + 7) >> 3

View File

@ -321,8 +321,8 @@ func TestMarshal(t *testing.T) {
t.Error(err) t.Error(err)
return return
} }
serialised := p224.Marshal(x, y) serialized := p224.Marshal(x, y)
xx, yy := p224.Unmarshal(serialised) xx, yy := p224.Unmarshal(serialized)
if xx == nil { if xx == nil {
t.Error("failed to unmarshal") t.Error("failed to unmarshal")
return return

View File

@ -190,7 +190,7 @@ func TestHMAC(t *testing.T) {
continue continue
} }
// Repetive Sum() calls should return the same value // Repetitive Sum() calls should return the same value
for k := 0; k < 2; k++ { for k := 0; k < 2; k++ {
sum := fmt.Sprintf("%x", h.Sum()) sum := fmt.Sprintf("%x", h.Sum())
if sum != tt.out { if sum != tt.out {

View File

@ -153,7 +153,7 @@ func (r *openpgpReader) Read(p []byte) (n int, err os.Error) {
// Decode reads a PGP armored block from the given Reader. It will ignore // Decode reads a PGP armored block from the given Reader. It will ignore
// leading garbage. If it doesn't find a block, it will return nil, os.EOF. The // leading garbage. If it doesn't find a block, it will return nil, os.EOF. The
// given Reader is not usable after calling this function: an arbitary amount // given Reader is not usable after calling this function: an arbitrary amount
// of data may have been read past the end of the block. // of data may have been read past the end of the block.
func Decode(in io.Reader) (p *Block, err os.Error) { func Decode(in io.Reader) (p *Block, err os.Error) {
r, _ := bufio.NewReaderSize(in, 100) r, _ := bufio.NewReaderSize(in, 100)

View File

@ -2,7 +2,7 @@
// Use of this source code is governed by a BSD-style // Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file. // license that can be found in the LICENSE file.
// Package packet implements parsing and serialisation of OpenPGP packets, as // Package packet implements parsing and serialization of OpenPGP packets, as
// specified in RFC 4880. // specified in RFC 4880.
package packet package packet
@ -386,7 +386,7 @@ func readMPI(r io.Reader) (mpi []byte, bitLength uint16, err os.Error) {
return return
} }
// mpiLength returns the length of the given *big.Int when serialised as an // mpiLength returns the length of the given *big.Int when serialized as an
// MPI. // MPI.
func mpiLength(n *big.Int) (mpiLengthInBytes int) { func mpiLength(n *big.Int) (mpiLengthInBytes int) {
mpiLengthInBytes = 2 /* MPI length */ mpiLengthInBytes = 2 /* MPI length */

View File

@ -293,7 +293,7 @@ type parsedMPI struct {
bitLength uint16 bitLength uint16
} }
// writeMPIs is a utility function for serialising several big integers to the // writeMPIs is a utility function for serializing several big integers to the
// given Writer. // given Writer.
func writeMPIs(w io.Writer, mpis ...parsedMPI) (err os.Error) { func writeMPIs(w io.Writer, mpis ...parsedMPI) (err os.Error) {
for _, mpi := range mpis { for _, mpi := range mpis {

View File

@ -393,7 +393,7 @@ func (sig *Signature) buildHashSuffix() (err os.Error) {
sig.HashSuffix[3], ok = s2k.HashToHashId(sig.Hash) sig.HashSuffix[3], ok = s2k.HashToHashId(sig.Hash)
if !ok { if !ok {
sig.HashSuffix = nil sig.HashSuffix = nil
return error.InvalidArgumentError("hash cannot be repesented in OpenPGP: " + strconv.Itoa(int(sig.Hash))) return error.InvalidArgumentError("hash cannot be represented in OpenPGP: " + strconv.Itoa(int(sig.Hash)))
} }
sig.HashSuffix[4] = byte(hashedSubpacketsLen >> 8) sig.HashSuffix[4] = byte(hashedSubpacketsLen >> 8)
sig.HashSuffix[5] = byte(hashedSubpacketsLen) sig.HashSuffix[5] = byte(hashedSubpacketsLen)

View File

@ -44,7 +44,7 @@ type MessageDetails struct {
DecryptedWith Key // the private key used to decrypt the message, if any. DecryptedWith Key // the private key used to decrypt the message, if any.
IsSigned bool // true if the message is signed. IsSigned bool // true if the message is signed.
SignedByKeyId uint64 // the key id of the signer, if any. SignedByKeyId uint64 // the key id of the signer, if any.
SignedBy *Key // the key of the signer, if availible. SignedBy *Key // the key of the signer, if available.
LiteralData *packet.LiteralData // the metadata of the contents LiteralData *packet.LiteralData // the metadata of the contents
UnverifiedBody io.Reader // the contents of the message. UnverifiedBody io.Reader // the contents of the message.
@ -145,7 +145,7 @@ ParsePackets:
// function so that it can decrypt a key or give us a passphrase. // function so that it can decrypt a key or give us a passphrase.
FindKey: FindKey:
for { for {
// See if any of the keys already have a private key availible // See if any of the keys already have a private key available
candidates = candidates[:0] candidates = candidates[:0]
candidateFingerprints := make(map[string]bool) candidateFingerprints := make(map[string]bool)
@ -214,7 +214,7 @@ FindKey:
return readSignedMessage(packets, md, keyring) return readSignedMessage(packets, md, keyring)
} }
// readSignedMessage reads a possibily signed message if mdin is non-zero then // readSignedMessage reads a possibly signed message if mdin is non-zero then
// that structure is updated and returned. Otherwise a fresh MessageDetails is // that structure is updated and returned. Otherwise a fresh MessageDetails is
// used. // used.
func readSignedMessage(packets *packet.Reader, mdin *MessageDetails, keyring KeyRing) (md *MessageDetails, err os.Error) { func readSignedMessage(packets *packet.Reader, mdin *MessageDetails, keyring KeyRing) (md *MessageDetails, err os.Error) {
@ -274,13 +274,13 @@ FindLiteralData:
// hashForSignature returns a pair of hashes that can be used to verify a // hashForSignature returns a pair of hashes that can be used to verify a
// signature. The signature may specify that the contents of the signed message // signature. The signature may specify that the contents of the signed message
// should be preprocessed (i.e. to normalise line endings). Thus this function // should be preprocessed (i.e. to normalize line endings). Thus this function
// returns two hashes. The second should be used to hash the message itself and // returns two hashes. The second should be used to hash the message itself and
// performs any needed preprocessing. // performs any needed preprocessing.
func hashForSignature(hashId crypto.Hash, sigType packet.SignatureType) (hash.Hash, hash.Hash, os.Error) { func hashForSignature(hashId crypto.Hash, sigType packet.SignatureType) (hash.Hash, hash.Hash, os.Error) {
h := hashId.New() h := hashId.New()
if h == nil { if h == nil {
return nil, nil, error.UnsupportedError("hash not availible: " + strconv.Itoa(int(hashId))) return nil, nil, error.UnsupportedError("hash not available: " + strconv.Itoa(int(hashId)))
} }
switch sigType { switch sigType {

View File

@ -193,9 +193,9 @@ func TestSymmetricallyEncrypted(t *testing.T) {
t.Errorf("ReadAll: %s", err) t.Errorf("ReadAll: %s", err)
} }
expectedCreatationTime := uint32(1295992998) expectedCreationTime := uint32(1295992998)
if md.LiteralData.Time != expectedCreatationTime { if md.LiteralData.Time != expectedCreationTime {
t.Errorf("LiteralData.Time is %d, want %d", md.LiteralData.Time, expectedCreatationTime) t.Errorf("LiteralData.Time is %d, want %d", md.LiteralData.Time, expectedCreationTime)
} }
if string(contents) != expected { if string(contents) != expected {

View File

@ -90,7 +90,7 @@ func Parse(r io.Reader) (f func(out, in []byte), err os.Error) {
} }
h := hash.New() h := hash.New()
if h == nil { if h == nil {
return nil, error.UnsupportedError("hash not availible: " + strconv.Itoa(int(hash))) return nil, error.UnsupportedError("hash not available: " + strconv.Itoa(int(hash)))
} }
switch buf[0] { switch buf[0] {

View File

@ -94,7 +94,7 @@ type PrivateKey struct {
Primes []*big.Int // prime factors of N, has >= 2 elements. Primes []*big.Int // prime factors of N, has >= 2 elements.
// Precomputed contains precomputed values that speed up private // Precomputed contains precomputed values that speed up private
// operations, if availible. // operations, if available.
Precomputed PrecomputedValues Precomputed PrecomputedValues
} }
@ -417,7 +417,7 @@ func decrypt(rand io.Reader, priv *PrivateKey, c *big.Int) (m *big.Int, err os.E
// Blinding enabled. Blinding involves multiplying c by r^e. // Blinding enabled. Blinding involves multiplying c by r^e.
// Then the decryption operation performs (m^e * r^e)^d mod n // Then the decryption operation performs (m^e * r^e)^d mod n
// which equals mr mod n. The factor of r can then be removed // which equals mr mod n. The factor of r can then be removed
// by multipling by the multiplicative inverse of r. // by multiplying by the multiplicative inverse of r.
var r *big.Int var r *big.Int

View File

@ -14,14 +14,14 @@ type TestConstantTimeCompareStruct struct {
out int out int
} }
var testConstandTimeCompareData = []TestConstantTimeCompareStruct{ var testConstantTimeCompareData = []TestConstantTimeCompareStruct{
{[]byte{}, []byte{}, 1}, {[]byte{}, []byte{}, 1},
{[]byte{0x11}, []byte{0x11}, 1}, {[]byte{0x11}, []byte{0x11}, 1},
{[]byte{0x12}, []byte{0x11}, 0}, {[]byte{0x12}, []byte{0x11}, 0},
} }
func TestConstantTimeCompare(t *testing.T) { func TestConstantTimeCompare(t *testing.T) {
for i, test := range testConstandTimeCompareData { for i, test := range testConstantTimeCompareData {
if r := ConstantTimeCompare(test.a, test.b); r != test.out { if r := ConstantTimeCompare(test.a, test.b); r != test.out {
t.Errorf("#%d bad result (got %x, want %x)", i, r, test.out) t.Errorf("#%d bad result (got %x, want %x)", i, r, test.out)
} }

View File

@ -87,7 +87,7 @@ const (
certTypeRSASign = 1 // A certificate containing an RSA key certTypeRSASign = 1 // A certificate containing an RSA key
certTypeDSSSign = 2 // A certificate containing a DSA key certTypeDSSSign = 2 // A certificate containing a DSA key
certTypeRSAFixedDH = 3 // A certificate containing a static DH key certTypeRSAFixedDH = 3 // A certificate containing a static DH key
certTypeDSSFixedDH = 4 // A certficiate containing a static DH key certTypeDSSFixedDH = 4 // A certificate containing a static DH key
// Rest of these are reserved by the TLS spec // Rest of these are reserved by the TLS spec
) )

View File

@ -34,7 +34,7 @@ type Conn struct {
cipherSuite uint16 cipherSuite uint16
ocspResponse []byte // stapled OCSP response ocspResponse []byte // stapled OCSP response
peerCertificates []*x509.Certificate peerCertificates []*x509.Certificate
// verifedChains contains the certificate chains that we built, as // verifiedChains contains the certificate chains that we built, as
// opposed to the ones presented by the server. // opposed to the ones presented by the server.
verifiedChains [][]*x509.Certificate verifiedChains [][]*x509.Certificate
@ -237,7 +237,7 @@ func (hc *halfConn) decrypt(b *block) (bool, alert) {
// "Password Interception in a SSL/TLS Channel", Brice // "Password Interception in a SSL/TLS Channel", Brice
// Canvel et al. // Canvel et al.
// //
// However, our behaviour matches OpenSSL, so we leak // However, our behavior matches OpenSSL, so we leak
// only as much as they do. // only as much as they do.
default: default:
panic("unknown cipher type") panic("unknown cipher type")
@ -410,7 +410,7 @@ func (hc *halfConn) freeBlock(b *block) {
// splitBlock splits a block after the first n bytes, // splitBlock splits a block after the first n bytes,
// returning a block with those n bytes and a // returning a block with those n bytes and a
// block with the remaindec. the latter may be nil. // block with the remainder. the latter may be nil.
func (hc *halfConn) splitBlock(b *block, n int) (*block, *block) { func (hc *halfConn) splitBlock(b *block, n int) (*block, *block) {
if len(b.data) <= n { if len(b.data) <= n {
return b, nil return b, nil

View File

@ -209,10 +209,10 @@ FindCipherSuite:
// If we received a client cert in response to our certificate request message, // If we received a client cert in response to our certificate request message,
// the client will send us a certificateVerifyMsg immediately after the // the client will send us a certificateVerifyMsg immediately after the
// clientKeyExchangeMsg. This message is a MD5SHA1 digest of all preceeding // clientKeyExchangeMsg. This message is a MD5SHA1 digest of all preceding
// handshake-layer messages that is signed using the private key corresponding // handshake-layer messages that is signed using the private key corresponding
// to the client's certificate. This allows us to verify that the client is in // to the client's certificate. This allows us to verify that the client is in
// posession of the private key of the certificate. // possession of the private key of the certificate.
if len(c.peerCertificates) > 0 { if len(c.peerCertificates) > 0 {
msg, err = c.readHandshake() msg, err = c.readHandshake()
if err != nil { if err != nil {

View File

@ -236,12 +236,12 @@ func (ka *ecdheRSAKeyAgreement) generateClientKeyExchange(config *Config, client
xBytes := x.Bytes() xBytes := x.Bytes()
copy(preMasterSecret[len(preMasterSecret)-len(xBytes):], xBytes) copy(preMasterSecret[len(preMasterSecret)-len(xBytes):], xBytes)
serialised := ka.curve.Marshal(mx, my) serialized := ka.curve.Marshal(mx, my)
ckx := new(clientKeyExchangeMsg) ckx := new(clientKeyExchangeMsg)
ckx.ciphertext = make([]byte, 1+len(serialised)) ckx.ciphertext = make([]byte, 1+len(serialized))
ckx.ciphertext[0] = byte(len(serialised)) ckx.ciphertext[0] = byte(len(serialized))
copy(ckx.ciphertext[1:], serialised) copy(ckx.ciphertext[1:], serialized)
return preMasterSecret, ckx, nil return preMasterSecret, ckx, nil
} }

View File

@ -33,10 +33,10 @@ type pkcs1PrivateKey struct {
Dq asn1.RawValue "optional" Dq asn1.RawValue "optional"
Qinv asn1.RawValue "optional" Qinv asn1.RawValue "optional"
AdditionalPrimes []pkcs1AddtionalRSAPrime "optional" AdditionalPrimes []pkcs1AdditionalRSAPrime "optional"
} }
type pkcs1AddtionalRSAPrime struct { type pkcs1AdditionalRSAPrime struct {
Prime asn1.RawValue Prime asn1.RawValue
// We ignore these values because rsa will calculate them. // We ignore these values because rsa will calculate them.
@ -135,7 +135,7 @@ func MarshalPKCS1PrivateKey(key *rsa.PrivateKey) []byte {
Qinv: rawValueForBig(key.Precomputed.Qinv), Qinv: rawValueForBig(key.Precomputed.Qinv),
} }
priv.AdditionalPrimes = make([]pkcs1AddtionalRSAPrime, len(key.Precomputed.CRTValues)) priv.AdditionalPrimes = make([]pkcs1AdditionalRSAPrime, len(key.Precomputed.CRTValues))
for i, values := range key.Precomputed.CRTValues { for i, values := range key.Precomputed.CRTValues {
priv.AdditionalPrimes[i].Prime = rawValueForBig(key.Primes[2+i]) priv.AdditionalPrimes[i].Prime = rawValueForBig(key.Primes[2+i])
priv.AdditionalPrimes[i].Exp = rawValueForBig(values.Exp) priv.AdditionalPrimes[i].Exp = rawValueForBig(values.Exp)
@ -280,7 +280,7 @@ var (
oidOrganizationalUnit = []int{2, 5, 4, 11} oidOrganizationalUnit = []int{2, 5, 4, 11}
oidCommonName = []int{2, 5, 4, 3} oidCommonName = []int{2, 5, 4, 3}
oidSerialNumber = []int{2, 5, 4, 5} oidSerialNumber = []int{2, 5, 4, 5}
oidLocatity = []int{2, 5, 4, 7} oidLocality = []int{2, 5, 4, 7}
oidProvince = []int{2, 5, 4, 8} oidProvince = []int{2, 5, 4, 8}
oidStreetAddress = []int{2, 5, 4, 9} oidStreetAddress = []int{2, 5, 4, 9}
oidPostalCode = []int{2, 5, 4, 17} oidPostalCode = []int{2, 5, 4, 17}
@ -308,7 +308,7 @@ func (n Name) toRDNSequence() (ret rdnSequence) {
ret = appendRDNs(ret, n.Country, oidCountry) ret = appendRDNs(ret, n.Country, oidCountry)
ret = appendRDNs(ret, n.Organization, oidOrganization) ret = appendRDNs(ret, n.Organization, oidOrganization)
ret = appendRDNs(ret, n.OrganizationalUnit, oidOrganizationalUnit) ret = appendRDNs(ret, n.OrganizationalUnit, oidOrganizationalUnit)
ret = appendRDNs(ret, n.Locality, oidLocatity) ret = appendRDNs(ret, n.Locality, oidLocality)
ret = appendRDNs(ret, n.Province, oidProvince) ret = appendRDNs(ret, n.Province, oidProvince)
ret = appendRDNs(ret, n.StreetAddress, oidStreetAddress) ret = appendRDNs(ret, n.StreetAddress, oidStreetAddress)
ret = appendRDNs(ret, n.PostalCode, oidPostalCode) ret = appendRDNs(ret, n.PostalCode, oidPostalCode)
@ -680,13 +680,13 @@ func parseCertificate(in *certificate) (*Certificate, os.Error) {
} }
case 19: case 19:
// RFC 5280, 4.2.1.9 // RFC 5280, 4.2.1.9
var constriants basicConstraints var constraints basicConstraints
_, err := asn1.Unmarshal(e.Value, &constriants) _, err := asn1.Unmarshal(e.Value, &constraints)
if err == nil { if err == nil {
out.BasicConstraintsValid = true out.BasicConstraintsValid = true
out.IsCA = constriants.IsCA out.IsCA = constraints.IsCA
out.MaxPathLen = constriants.MaxPathLen out.MaxPathLen = constraints.MaxPathLen
continue continue
} }
case 17: case 17:

View File

@ -22,7 +22,7 @@ func blockToUint32(src []byte) (uint32, uint32) {
return r0, r1 return r0, r1
} }
// uint32ToBlock writes two unint32s into an 8 byte data block. // uint32ToBlock writes two uint32s into an 8 byte data block.
// Values are written as big endian. // Values are written as big endian.
func uint32ToBlock(v0, v1 uint32, dst []byte) { func uint32ToBlock(v0, v1 uint32, dst []byte) {
dst[0] = byte(v0 >> 24) dst[0] = byte(v0 >> 24)

View File

@ -8,7 +8,7 @@ import (
"testing" "testing"
) )
// A sample test key for when we just want to initialise a cipher // A sample test key for when we just want to initialize a cipher
var testKey = []byte{0x00, 0x11, 0x22, 0x33, 0x44, 0x55, 0x66, 0x77, 0x88, 0x99, 0xAA, 0xBB, 0xCC, 0xDD, 0xEE, 0xFF} var testKey = []byte{0x00, 0x11, 0x22, 0x33, 0x44, 0x55, 0x66, 0x77, 0x88, 0x99, 0xAA, 0xBB, 0xCC, 0xDD, 0xEE, 0xFF}
// Test that the block size for XTEA is correct // Test that the block size for XTEA is correct
@ -26,12 +26,12 @@ func TestBlocksize(t *testing.T) {
result := c.BlockSize() result := c.BlockSize()
if result != 8 { if result != 8 {
t.Errorf("BlockSize function - expected 8, gotr %d", result) t.Errorf("BlockSize function - expected 8, got %d", result)
return return
} }
} }
// A series of test values to confirm that the Cipher.table array was initialised correctly // A series of test values to confirm that the Cipher.table array was initialized correctly
var testTable = []uint32{ var testTable = []uint32{
0x00112233, 0x6B1568B8, 0xE28CE030, 0xC5089E2D, 0xC5089E2D, 0x1EFBD3A2, 0xA7845C2A, 0x78EF0917, 0x00112233, 0x6B1568B8, 0xE28CE030, 0xC5089E2D, 0xC5089E2D, 0x1EFBD3A2, 0xA7845C2A, 0x78EF0917,
0x78EF0917, 0x172682D0, 0x5B6AC714, 0x822AC955, 0x3DE68511, 0xDC1DFECA, 0x2062430E, 0x3611343F, 0x78EF0917, 0x172682D0, 0x5B6AC714, 0x822AC955, 0x3DE68511, 0xDC1DFECA, 0x2062430E, 0x3611343F,
@ -43,7 +43,7 @@ var testTable = []uint32{
0x4E22726F, 0x309E306C, 0x309E306C, 0x8A9165E1, 0x1319EE69, 0xF595AC66, 0xF595AC66, 0x4F88E1DB, 0x4E22726F, 0x309E306C, 0x309E306C, 0x8A9165E1, 0x1319EE69, 0xF595AC66, 0xF595AC66, 0x4F88E1DB,
} }
// Test that the cipher context is initialised correctly // Test that the cipher context is initialized correctly
func TestCipherInit(t *testing.T) { func TestCipherInit(t *testing.T) {
c, err := NewCipher(testKey) c, err := NewCipher(testKey)
if err != nil { if err != nil {
@ -53,7 +53,7 @@ func TestCipherInit(t *testing.T) {
for i := 0; i < len(c.table); i++ { for i := 0; i < len(c.table); i++ {
if c.table[i] != testTable[i] { if c.table[i] != testTable[i] {
t.Errorf("NewCipher() failed to initialise Cipher.table[%d] correctly. Expected %08X, got %08X", i, testTable[i], c.table[i]) t.Errorf("NewCipher() failed to initialize Cipher.table[%d] correctly. Expected %08X, got %08X", i, testTable[i], c.table[i])
break break
} }
} }

View File

@ -523,7 +523,7 @@ func (d *Data) Type(off Offset) (Type, os.Error) {
// Attributes: // Attributes:
// AttrType: type of return value if any // AttrType: type of return value if any
// AttrName: possible name of type [ignored] // AttrName: possible name of type [ignored]
// AttrPrototyped: whether used ANSI C prototye [ignored] // AttrPrototyped: whether used ANSI C prototype [ignored]
// Children: // Children:
// TagFormalParameter: typed parameter // TagFormalParameter: typed parameter
// AttrType: type of parameter // AttrType: type of parameter

View File

@ -184,7 +184,7 @@ func (f *File) Close() os.Error {
return err return err
} }
// NewFile creates a new File for acecssing a Mach-O binary in an underlying reader. // NewFile creates a new File for accessing a Mach-O binary in an underlying reader.
// The Mach-O binary is expected to start at position 0 in the ReaderAt. // The Mach-O binary is expected to start at position 0 in the ReaderAt.
func NewFile(r io.ReaderAt) (*File, os.Error) { func NewFile(r io.ReaderAt) (*File, os.Error) {
f := new(File) f := new(File)

View File

@ -112,7 +112,7 @@ func (f *File) Close() os.Error {
return err return err
} }
// NewFile creates a new File for acecssing a PE binary in an underlying reader. // NewFile creates a new File for accessing a PE binary in an underlying reader.
func NewFile(r io.ReaderAt) (*File, os.Error) { func NewFile(r io.ReaderAt) (*File, os.Error) {
f := new(File) f := new(File)
sr := io.NewSectionReader(r, 0, 1<<63-1) sr := io.NewSectionReader(r, 0, 1<<63-1)

View File

@ -277,7 +277,7 @@ func (t *thread) ptraceDetach() os.Error {
} }
/* /*
* Logging utilties * Logging utilities
*/ */
var logLock sync.Mutex var logLock sync.Mutex
@ -1192,7 +1192,7 @@ func (p *process) attachAllThreads() os.Error {
// We stop threads as we attach to them; however, because new // We stop threads as we attach to them; however, because new
// threads can appear while we're looping over all of them, we // threads can appear while we're looping over all of them, we
// have to repeatly scan until we know we're attached to all // have to repeatedly scan until we know we're attached to all
// of them. // of them.
for again := true; again; { for again := true; again; {
again = false again = false
@ -1214,7 +1214,7 @@ func (p *process) attachAllThreads() os.Error {
_, err = p.attachThread(tid) _, err = p.attachThread(tid)
if err != nil { if err != nil {
// There could have been a race, or // There could have been a race, or
// this process could be a zobmie. // this process could be a zombie.
statFile, err2 := ioutil.ReadFile(taskPath + "/" + tidStr + "/stat") statFile, err2 := ioutil.ReadFile(taskPath + "/" + tidStr + "/stat")
if err2 != nil { if err2 != nil {
switch err2 := err2.(type) { switch err2 := err2.(type) {

View File

@ -273,5 +273,5 @@ func (d *decoder) Read(p []byte) (n int, err os.Error) {
d.nbuf = copy(d.buf[0:], d.buf[nl+1:d.nbuf]) d.nbuf = copy(d.buf[0:], d.buf[nl+1:d.nbuf])
d.off += int64(nl + 1) d.off += int64(nl + 1)
} }
panic("unreacahable") panic("unreachable")
} }

View File

@ -140,7 +140,7 @@ Error:
// any lines which could be header lines. However, a valid preamble // any lines which could be header lines. However, a valid preamble
// line is not a valid header line, therefore we cannot have consumed // line is not a valid header line, therefore we cannot have consumed
// the preamble line for the any subsequent block. Thus, we will always // the preamble line for the any subsequent block. Thus, we will always
// find any valid block, no matter what bytes preceed it. // find any valid block, no matter what bytes precede it.
// //
// For example, if the input is // For example, if the input is
// //

View File

@ -594,7 +594,7 @@ func (s *State) eval(fexpr expr, value reflect.Value, index int) bool {
s.eval(t.indent, value, index) s.eval(t.indent, value, index)
// if the indentation evaluates to nil, the state's output buffer // if the indentation evaluates to nil, the state's output buffer
// didn't change - either way it's ok to append the difference to // didn't change - either way it's ok to append the difference to
// the current identation // the current indentation
s.indent.Write(s.output.Bytes()[mark.outputLen:s.output.Len()]) s.indent.Write(s.output.Bytes()[mark.outputLen:s.output.Len()])
s.restore(mark) s.restore(mark)

View File

@ -310,7 +310,7 @@ func authenticate(w *bufio.Writer, displayStr string) os.Error {
return os.NewError("unsupported Xauth") return os.NewError("unsupported Xauth")
} }
// 0x006c means little-endian. 0x000b, 0x0000 means X major version 11, minor version 0. // 0x006c means little-endian. 0x000b, 0x0000 means X major version 11, minor version 0.
// 0x0012 and 0x0010 means the auth key and value have lenths 18 and 16. // 0x0012 and 0x0010 means the auth key and value have lengths 18 and 16.
// The final 0x0000 is padding, so that the string length is a multiple of 4. // The final 0x0000 is padding, so that the string length is a multiple of 4.
_, err = io.WriteString(w, "\x6c\x00\x0b\x00\x00\x00\x12\x00\x10\x00\x00\x00") _, err = io.WriteString(w, "\x6c\x00\x0b\x00\x00\x00\x12\x00\x10\x00\x00\x00")
if err != nil { if err != nil {
@ -517,7 +517,7 @@ func (c *conn) handshake() os.Error {
if err != nil { if err != nil {
return err return err
} }
// Ignore some things that we don't care about (totalling 10 + vendorLen bytes): // Ignore some things that we don't care about (totaling 10 + vendorLen bytes):
// imageByteOrder(1), bitmapFormatBitOrder(1), bitmapFormatScanlineUnit(1) bitmapFormatScanlinePad(1), // imageByteOrder(1), bitmapFormatBitOrder(1), bitmapFormatScanlineUnit(1) bitmapFormatScanlinePad(1),
// minKeycode(1), maxKeycode(1), padding(4), vendor (vendorLen). // minKeycode(1), maxKeycode(1), padding(4), vendor (vendorLen).
if 10+int(vendorLen) > cap(c.buf) { if 10+int(vendorLen) > cap(c.buf) {

View File

@ -1781,7 +1781,7 @@ func (a *exprInfo) compileBinaryExpr(op token.Token, l, r *expr) *expr {
// written: Function values are equal if they were // written: Function values are equal if they were
// created by the same execution of a function literal // created by the same execution of a function literal
// or refer to the same function declaration. This is // or refer to the same function declaration. This is
// *almost* but not quite waht 6g implements. If a // *almost* but not quite what 6g implements. If a
// function literals does not capture any variables, // function literals does not capture any variables,
// then multiple executions of it will result in the // then multiple executions of it will result in the
// same closure. Russ says he'll change that. // same closure. Russ says he'll change that.

View File

@ -68,7 +68,7 @@ type flowBuf struct {
gotos map[token.Pos]*flowBlock gotos map[token.Pos]*flowBlock
// labels is a map from label name to information on the block // labels is a map from label name to information on the block
// at the point of the label. labels are tracked by name, // at the point of the label. labels are tracked by name,
// since mutliple labels at the same PC can have different // since multiple labels at the same PC can have different
// blocks. // blocks.
labels map[string]*flowBlock labels map[string]*flowBlock
} }
@ -307,7 +307,7 @@ func (a *stmtCompiler) compile(s ast.Stmt) {
} }
if notimpl { if notimpl {
a.diag("%T statment node not implemented", s) a.diag("%T statement node not implemented", s)
} }
if a.block.inner != nil { if a.block.inner != nil {
@ -550,7 +550,7 @@ func (a *stmtCompiler) doAssign(lhs []ast.Expr, rhs []ast.Expr, tok token.Token,
ident, ok = le.(*ast.Ident) ident, ok = le.(*ast.Ident)
if !ok { if !ok {
a.diagAt(le.Pos(), "left side of := must be a name") a.diagAt(le.Pos(), "left side of := must be a name")
// Suppress new defitions errors // Suppress new definitions errors
nDefs++ nDefs++
continue continue
} }

View File

@ -68,7 +68,7 @@ func (a *typeCompiler) compileArrayType(x *ast.ArrayType, allowRec bool) Type {
} }
if _, ok := x.Len.(*ast.Ellipsis); ok { if _, ok := x.Len.(*ast.Ellipsis); ok {
a.diagAt(x.Len.Pos(), "... array initailizers not implemented") a.diagAt(x.Len.Pos(), "... array initializers not implemented")
return nil return nil
} }
l, ok := a.compileArrayLen(a.block, x.Len) l, ok := a.compileArrayLen(a.block, x.Len)

View File

@ -96,7 +96,7 @@ const (
// Some button control styles // Some button control styles
BS_DEFPUSHBUTTON = 1 BS_DEFPUSHBUTTON = 1
// Some colour constants // Some color constants
COLOR_WINDOW = 5 COLOR_WINDOW = 5
COLOR_BTNFACE = 15 COLOR_BTNFACE = 15
@ -108,13 +108,13 @@ const (
) )
var ( var (
// Some globaly known cusrors // Some globally known cursors
IDC_ARROW = MakeIntResource(32512) IDC_ARROW = MakeIntResource(32512)
IDC_IBEAM = MakeIntResource(32513) IDC_IBEAM = MakeIntResource(32513)
IDC_WAIT = MakeIntResource(32514) IDC_WAIT = MakeIntResource(32514)
IDC_CROSS = MakeIntResource(32515) IDC_CROSS = MakeIntResource(32515)
// Some globaly known icons // Some globally known icons
IDI_APPLICATION = MakeIntResource(32512) IDI_APPLICATION = MakeIntResource(32512)
IDI_HAND = MakeIntResource(32513) IDI_HAND = MakeIntResource(32513)
IDI_QUESTION = MakeIntResource(32514) IDI_QUESTION = MakeIntResource(32514)

View File

@ -76,33 +76,33 @@ func TestString(t *testing.T) {
} }
func TestMapCounter(t *testing.T) { func TestMapCounter(t *testing.T) {
colours := NewMap("bike-shed-colours") colors := NewMap("bike-shed-colors")
colours.Add("red", 1) colors.Add("red", 1)
colours.Add("red", 2) colors.Add("red", 2)
colours.Add("blue", 4) colors.Add("blue", 4)
colours.AddFloat("green", 4.125) colors.AddFloat("green", 4.125)
if x := colours.m["red"].(*Int).i; x != 3 { if x := colors.m["red"].(*Int).i; x != 3 {
t.Errorf("colours.m[\"red\"] = %v, want 3", x) t.Errorf("colors.m[\"red\"] = %v, want 3", x)
} }
if x := colours.m["blue"].(*Int).i; x != 4 { if x := colors.m["blue"].(*Int).i; x != 4 {
t.Errorf("colours.m[\"blue\"] = %v, want 4", x) t.Errorf("colors.m[\"blue\"] = %v, want 4", x)
} }
if x := colours.m["green"].(*Float).f; x != 4.125 { if x := colors.m["green"].(*Float).f; x != 4.125 {
t.Errorf("colours.m[\"green\"] = %v, want 3.14", x) t.Errorf("colors.m[\"green\"] = %v, want 3.14", x)
} }
// colours.String() should be '{"red":3, "blue":4}', // colors.String() should be '{"red":3, "blue":4}',
// though the order of red and blue could vary. // though the order of red and blue could vary.
s := colours.String() s := colors.String()
var j interface{} var j interface{}
err := json.Unmarshal([]byte(s), &j) err := json.Unmarshal([]byte(s), &j)
if err != nil { if err != nil {
t.Errorf("colours.String() isn't valid JSON: %v", err) t.Errorf("colors.String() isn't valid JSON: %v", err)
} }
m, ok := j.(map[string]interface{}) m, ok := j.(map[string]interface{})
if !ok { if !ok {
t.Error("colours.String() didn't produce a map.") t.Error("colors.String() didn't produce a map.")
} }
red := m["red"] red := m["red"]
x, ok := red.(float64) x, ok := red.(float64)

View File

@ -515,10 +515,10 @@ type (
// An EmptyStmt node represents an empty statement. // An EmptyStmt node represents an empty statement.
// The "position" of the empty statement is the position // The "position" of the empty statement is the position
// of the immediately preceeding semicolon. // of the immediately preceding semicolon.
// //
EmptyStmt struct { EmptyStmt struct {
Semicolon token.Pos // position of preceeding ";" Semicolon token.Pos // position of preceding ";"
} }
// A LabeledStmt node represents a labeled statement. // A LabeledStmt node represents a labeled statement.
@ -596,7 +596,7 @@ type (
// An IfStmt node represents an if statement. // An IfStmt node represents an if statement.
IfStmt struct { IfStmt struct {
If token.Pos // position of "if" keyword If token.Pos // position of "if" keyword
Init Stmt // initalization statement; or nil Init Stmt // initialization statement; or nil
Cond Expr // condition Cond Expr // condition
Body *BlockStmt Body *BlockStmt
Else Stmt // else branch; or nil Else Stmt // else branch; or nil
@ -613,7 +613,7 @@ type (
// A SwitchStmt node represents an expression switch statement. // A SwitchStmt node represents an expression switch statement.
SwitchStmt struct { SwitchStmt struct {
Switch token.Pos // position of "switch" keyword Switch token.Pos // position of "switch" keyword
Init Stmt // initalization statement; or nil Init Stmt // initialization statement; or nil
Tag Expr // tag expression; or nil Tag Expr // tag expression; or nil
Body *BlockStmt // CaseClauses only Body *BlockStmt // CaseClauses only
} }
@ -621,7 +621,7 @@ type (
// An TypeSwitchStmt node represents a type switch statement. // An TypeSwitchStmt node represents a type switch statement.
TypeSwitchStmt struct { TypeSwitchStmt struct {
Switch token.Pos // position of "switch" keyword Switch token.Pos // position of "switch" keyword
Init Stmt // initalization statement; or nil Init Stmt // initialization statement; or nil
Assign Stmt // x := y.(type) or y.(type) Assign Stmt // x := y.(type) or y.(type)
Body *BlockStmt // CaseClauses only Body *BlockStmt // CaseClauses only
} }
@ -643,7 +643,7 @@ type (
// A ForStmt represents a for statement. // A ForStmt represents a for statement.
ForStmt struct { ForStmt struct {
For token.Pos // position of "for" keyword For token.Pos // position of "for" keyword
Init Stmt // initalization statement; or nil Init Stmt // initialization statement; or nil
Cond Expr // condition; or nil Cond Expr // condition; or nil
Post Stmt // post iteration statement; or nil Post Stmt // post iteration statement; or nil
Body *BlockStmt Body *BlockStmt

View File

@ -54,7 +54,7 @@ type parser struct {
// Non-syntactic parser control // Non-syntactic parser control
exprLev int // < 0: in control clause, >= 0: in expression exprLev int // < 0: in control clause, >= 0: in expression
// Ordinary identifer scopes // Ordinary identifier scopes
pkgScope *ast.Scope // pkgScope.Outer == nil pkgScope *ast.Scope // pkgScope.Outer == nil
topScope *ast.Scope // top-most scope; may be pkgScope topScope *ast.Scope // top-most scope; may be pkgScope
unresolved []*ast.Ident // unresolved identifiers unresolved []*ast.Ident // unresolved identifiers

View File

@ -33,7 +33,7 @@ import (
// line break was printed; returns false otherwise. // line break was printed; returns false otherwise.
// //
// TODO(gri): linebreak may add too many lines if the next statement at "line" // TODO(gri): linebreak may add too many lines if the next statement at "line"
// is preceeded by comments because the computation of n assumes // is preceded by comments because the computation of n assumes
// the current position before the comment and the target position // the current position before the comment and the target position
// after the comment. Thus, after interspersing such comments, the // after the comment. Thus, after interspersing such comments, the
// space taken up by them is not considered to reduce the number of // space taken up by them is not considered to reduce the number of
@ -438,7 +438,7 @@ func (p *printer) fieldList(fields *ast.FieldList, isStruct, isIncomplete bool)
if len(list) > 0 { if len(list) > 0 {
p.print(formfeed) p.print(formfeed)
} }
p.flush(p.fset.Position(rbrace), token.RBRACE) // make sure we don't loose the last line comment p.flush(p.fset.Position(rbrace), token.RBRACE) // make sure we don't lose the last line comment
p.setLineComment("// contains filtered or unexported fields") p.setLineComment("// contains filtered or unexported fields")
} }
@ -465,7 +465,7 @@ func (p *printer) fieldList(fields *ast.FieldList, isStruct, isIncomplete bool)
if len(list) > 0 { if len(list) > 0 {
p.print(formfeed) p.print(formfeed)
} }
p.flush(p.fset.Position(rbrace), token.RBRACE) // make sure we don't loose the last line comment p.flush(p.fset.Position(rbrace), token.RBRACE) // make sure we don't lose the last line comment
p.setLineComment("// contains filtered or unexported methods") p.setLineComment("// contains filtered or unexported methods")
} }
@ -1390,7 +1390,7 @@ func (p *printer) nodeSize(n ast.Node, maxSize int) (size int) {
size = maxSize + 1 // assume n doesn't fit size = maxSize + 1 // assume n doesn't fit
p.nodeSizes[n] = size p.nodeSizes[n] = size
// nodeSize computation must be indendent of particular // nodeSize computation must be independent of particular
// style so that we always get the same decision; print // style so that we always get the same decision; print
// in RawFormat // in RawFormat
cfg := Config{Mode: RawFormat} cfg := Config{Mode: RawFormat}

View File

@ -589,7 +589,7 @@ func (p *printer) writeCommentSuffix(needsLinebreak bool) (droppedFF bool) {
// ignore trailing whitespace // ignore trailing whitespace
p.wsbuf[i] = ignore p.wsbuf[i] = ignore
case indent, unindent: case indent, unindent:
// don't loose indentation information // don't lose indentation information
case newline, formfeed: case newline, formfeed:
// if we need a line break, keep exactly one // if we need a line break, keep exactly one
// but remember if we dropped any formfeeds // but remember if we dropped any formfeeds

View File

@ -436,7 +436,7 @@ func _() {
// Comments immediately adjacent to punctuation (for which the go/printer // Comments immediately adjacent to punctuation (for which the go/printer
// may obly have estimated position information) must remain after the punctuation. // may only have estimated position information) must remain after the punctuation.
func _() { func _() {
_ = T{ _ = T{
1, // comment after comma 1, // comment after comma

View File

@ -434,7 +434,7 @@ func _() {
// Comments immediately adjacent to punctuation (for which the go/printer // Comments immediately adjacent to punctuation (for which the go/printer
// may obly have estimated position information) must remain after the punctuation. // may only have estimated position information) must remain after the punctuation.
func _() { func _() {
_ = T{ _ = T{
1, // comment after comma 1, // comment after comma

View File

@ -89,7 +89,7 @@ var tokens = [...]elt{
literal, literal,
}, },
// Operators and delimitors // Operators and delimiters
{token.ADD, "+", operator}, {token.ADD, "+", operator},
{token.SUB, "-", operator}, {token.SUB, "-", operator},
{token.MUL, "*", operator}, {token.MUL, "*", operator},

View File

@ -135,7 +135,7 @@ func (f *File) position(p Pos) (pos Position) {
func (s *FileSet) Position(p Pos) (pos Position) { func (s *FileSet) Position(p Pos) (pos Position) {
if p != NoPos { if p != NoPos {
// TODO(gri) consider optimizing the case where p // TODO(gri) consider optimizing the case where p
// is in the last file addded, or perhaps // is in the last file added, or perhaps
// looked at - will eliminate one level // looked at - will eliminate one level
// of search // of search
s.mutex.RLock() s.mutex.RLock()

View File

@ -172,7 +172,7 @@ func ignoreTwoUints(i *decInstr, state *decoderState, p unsafe.Pointer) {
state.decodeUint() state.decodeUint()
} }
// decBool decodes a uiint and stores it as a boolean through p. // decBool decodes a uint and stores it as a boolean through p.
func decBool(i *decInstr, state *decoderState, p unsafe.Pointer) { func decBool(i *decInstr, state *decoderState, p unsafe.Pointer) {
if i.indir > 0 { if i.indir > 0 {
if *(*unsafe.Pointer)(p) == nil { if *(*unsafe.Pointer)(p) == nil {

View File

@ -384,7 +384,7 @@ func TestGobEncoderFieldTypeError(t *testing.T) {
y := &GobTest1{} y := &GobTest1{}
err = dec.Decode(y) err = dec.Decode(y)
if err == nil { if err == nil {
t.Fatal("expected decode error for mistmatched fields (non-encoder to decoder)") t.Fatal("expected decode error for mismatched fields (non-encoder to decoder)")
} }
if strings.Index(err.String(), "type") < 0 { if strings.Index(err.String(), "type") < 0 {
t.Fatal("expected type error; got", err) t.Fatal("expected type error; got", err)

View File

@ -673,7 +673,7 @@ func mustGetTypeInfo(rt reflect.Type) *typeInfo {
// A type that implements GobEncoder and GobDecoder has complete // A type that implements GobEncoder and GobDecoder has complete
// control over the representation of its data and may therefore // control over the representation of its data and may therefore
// contain things such as private fields, channels, and functions, // contain things such as private fields, channels, and functions,
// which are not usually transmissable in gob streams. // which are not usually transmissible in gob streams.
// //
// Note: Since gobs can be stored permanently, It is good design // Note: Since gobs can be stored permanently, It is good design
// to guarantee the encoding used by a GobEncoder is stable as the // to guarantee the encoding used by a GobEncoder is stable as the

View File

@ -100,7 +100,7 @@ var tokenTests = []tokenTest{
"<p \t\n iD=\"a&quot;B\" foo=\"bar\"><EM>te&lt;&amp;;xt</em></p>", "<p \t\n iD=\"a&quot;B\" foo=\"bar\"><EM>te&lt;&amp;;xt</em></p>",
`<p id="a&quot;B" foo="bar">$<em>$te&lt;&amp;;xt$</em>$</p>`, `<p id="a&quot;B" foo="bar">$<em>$te&lt;&amp;;xt$</em>$</p>`,
}, },
// A non-existant entity. Tokenizing and converting back to a string should // A nonexistent entity. Tokenizing and converting back to a string should
// escape the "&" to become "&amp;". // escape the "&" to become "&amp;".
{ {
"noSuchEntity", "noSuchEntity",

View File

@ -18,7 +18,7 @@ func NewChunkedWriter(w io.Writer) io.WriteCloser {
} }
// Writing to ChunkedWriter translates to writing in HTTP chunked Transfer // Writing to ChunkedWriter translates to writing in HTTP chunked Transfer
// Encoding wire format to the undering Wire writer. // Encoding wire format to the underlying Wire writer.
type chunkedWriter struct { type chunkedWriter struct {
Wire io.Writer Wire io.Writer
} }

View File

@ -111,7 +111,7 @@ func (sc *ServerConn) Read() (req *Request, err os.Error) {
// Make sure body is fully consumed, even if user does not call body.Close // Make sure body is fully consumed, even if user does not call body.Close
if lastbody != nil { if lastbody != nil {
// body.Close is assumed to be idempotent and multiple calls to // body.Close is assumed to be idempotent and multiple calls to
// it should return the error that its first invokation // it should return the error that its first invocation
// returned. // returned.
err = lastbody.Close() err = lastbody.Close()
if err != nil { if err != nil {

View File

@ -551,7 +551,7 @@ var serverExpectTests = []serverExpectTest{
{100, "", true, "200 OK"}, {100, "", true, "200 OK"},
// 100-continue but requesting client to deny us, // 100-continue but requesting client to deny us,
// so it never eads the body. // so it never reads the body.
{100, "100-continue", false, "401 Unauthorized"}, {100, "100-continue", false, "401 Unauthorized"},
// Likewise without 100-continue: // Likewise without 100-continue:
{100, "", false, "401 Unauthorized"}, {100, "", false, "401 Unauthorized"},