yesman82
yesman82

Reputation: 451

Print unicode character from variable (swift)

I have a problem I couldn't find a solution to. I have a string variable holding the unicode "1f44d" and I want to convert it to a unicode character 👍.

Usually one would do something like this:

println("\u{1f44d}") // 👍

Here is what I mean:

let charAsString = "1f44d" // code in variable
println("\u{\(charAsString)}") // not working

I have tried several other ways but somehow the workings behind this magic stay hidden for me.

One should imagine the value of charAsString coming from an API call or from another object.

Upvotes: 19

Views: 15370

Answers (7)

Joule87
Joule87

Reputation: 651

I made this extension that works pretty well:

extension String {
    var unicode: String? {
        if let charCode = UInt32(self, radix: 16),
           let unicode = UnicodeScalar(charCode) {
            let str = String(unicode)
            return str
        }
        return nil
    }
}

How to test it:

if let test = "e9c8".unicode {
    print(test)
}

//print: enter image description here

Upvotes: 1

hkdalex
hkdalex

Reputation: 755

Here are a couple ways to do it:

let string = "1f44d"

Solution 1:

"&#x\(string);".applyingTransform(.toXMLHex, reverse: true)

Solution 2:

"U+\(string)".applyingTransform(StringTransform("Hex/Unicode"), reverse: true)

Upvotes: 1

Imanou Petit
Imanou Petit

Reputation: 92599

You cannot use string interpolation in Swift as you try to use it. Therefore, the following code won't compile:

let charAsString = "1f44d"
print("\u{\(charAsString)}")

You will have to convert your string variable into an integer (using init(_:radix:) initializer) then create a Unicode scalar from this integer. The Swift 5 Playground sample code below shows how to proceed:

let validCodeString = "1f44d"
let validUnicodeScalarValue = Int(validCodeString, radix: 16)!
let validUnicodeScalar = Unicode.Scalar(validUnicodeScalarValue)!
print(validUnicodeScalar) // 👍

Upvotes: 0

Avinash
Avinash

Reputation: 4372

You can use

let char = "-12"
print(char.unicodeScalars.map {$0.value }))

You'll get the values as:

[45, 49, 50]

Upvotes: 1

Fantattitude
Fantattitude

Reputation: 1842

As of Swift 2.0, every Int type has an initializer able to take String as an input. You can then easily generate an UnicodeScalar corresponding and print it afterwards. Without having to change your representation of chars as string ;).

UPDATED: Swift 3.0 changed UnicodeScalar initializer

print("\u{1f44d}") // 👍

let charAsString = "1f44d" // code in variable

let charAsInt = Int(charAsString, radix: 16)! // As indicated by @MartinR radix is required, default won't do it
let uScalar = UnicodeScalar(charAsInt)! // In Swift 3.0 this initializer is failible so you'll need either force unwrap or optionnal unwrapping

print("\(uScalar)")

Upvotes: 7

Martin R
Martin R

Reputation: 540105

One possible solution (explanations "inline"):

let charAsString = "1f44d"

// Convert hex string to numeric value first:
var charCode : UInt32 = 0
let scanner = NSScanner(string: charAsString)
if scanner.scanHexInt(&charCode) {

    // Create string from Unicode code point:
    let str = String(UnicodeScalar(charCode))
    println(str) // 👍
} else {
    println("invalid input")
}

Slightly simpler with Swift 2:

let charAsString = "1f44d"

// Convert hex string to numeric value first:
if let charCode = UInt32(charAsString, radix: 16) {
    // Create string from Unicode code point:
    let str = String(UnicodeScalar(charCode))
    print(str) // 👍
} else {
    print("invalid input")
}

Note also that not all code points are valid Unicode scalars, compare Validate Unicode code point in Swift.


Update for Swift 3:

public init?(_ v: UInt32)

is now a failable initializer of UnicodeScalar and checks if the given numeric input is a valid Unicode scalar value:

let charAsString = "1f44d"

// Convert hex string to numeric value first:
if let charCode = UInt32(charAsString, radix: 16),
    let unicode = UnicodeScalar(charCode) {
    // Create string from Unicode code point:
    let str = String(unicode)
    print(str) // 👍
} else {
    print("invalid input")
}

Upvotes: 14

Jakub Vano
Jakub Vano

Reputation: 3873

This can be done in two steps:

  1. convert charAsString to Int code
  2. convert code to unicode character

Second step can be done e.g. like this

var code = 0x1f44d
var scalar = UnicodeScalar(code)
var string = "\(scalar)"

As for first the step, see here how to convert String in hex representation to Int

Upvotes: 12

Related Questions