code
stringlengths 5
1.03M
| repo_name
stringlengths 5
90
| path
stringlengths 4
158
| license
stringclasses 15
values | size
int64 5
1.03M
| n_ast_errors
int64 0
53.9k
| ast_max_depth
int64 2
4.17k
| n_whitespaces
int64 0
365k
| n_ast_nodes
int64 3
317k
| n_ast_terminals
int64 1
171k
| n_ast_nonterminals
int64 1
146k
| loc
int64 -1
37.3k
| cycloplexity
int64 -1
1.31k
|
---|---|---|---|---|---|---|---|---|---|---|---|---|
{-# LANGUAGE PolyKinds #-}
{-# LANGUAGE FlexibleContexts #-}
{-# LANGUAGE DataKinds, KindSignatures,GADTs,
TypeOperators,
FlexibleInstances,
ScopedTypeVariables #-}
module Data.Union (
Union (..),
inject, project, wrap, unwrap,
push, pop,
swap, rotate,
conceal
) where
import Data.Type.Equality ((:~:) (..), gcastWith, testEquality)
import Data.Type.List hiding (length)
import Data.Type.List.Index (Index (..))
import qualified Data.Type.List.Index as Index
import Algebra.Absurd
----------------------------------------------------------------------------
data Union (l :: [k]) where
Union :: Index l e -> e -> Union l
instance Absurd (Union '[]) where
absurd (Union x _) = absurd x
inject :: Member e l => e -> Union l
inject = Union Index.index
project :: forall l e. Member e l => Union l -> Maybe e
project (Union i x) = fmap (\refl -> gcastWith refl x) $ testEquality i (Index.index :: Index l e)
wrap :: e -> Union (e ': l)
wrap = inject
unwrap :: Union '[e] -> e
unwrap (Union i x) = gcastWith (Index.trivial i) x
----------------------------------------------------------------------------
push :: Union l -> Union (e ': l)
push (Union i x) = Union (Index.push i) x
pop :: Union (e ': l) -> Either e (Union l)
pop (Union i x) = case Index.pop i of
Left Refl -> Left x
Right i' -> Right $ Union i' x
swap :: Union (e ': f ': l) -> Union (f ': e ': l)
swap (Union i x) = Union (Index.swap i) x
rotate :: Union (e ': f ': g ': l) -> Union (f ': g ': e ': l)
rotate (Union i x) = Union (Index.rotate i) x
----------------------------------------------------------------------------
conceal :: Member e l => Union (e ': l) -> Union l
conceal (Union i x) = Union (Index.conceal i) x
| seagull-kamome/haskell-toybox | Data/Union.hs | bsd-3-clause | 1,798 | 0 | 10 | 393 | 735 | 389 | 346 | 41 | 2 |
{-# OPTIONS_GHC -Wall #-}
module Main where
import Graphics.Rendering.Chart hiding (c)
import Graphics.Rendering.Chart.Gtk
import Data.Accessor
import Text.Printf
import Design.Config
import Design.WorkingConfig
import Aero.Drag.WettedArea
import Aero.Drag.Upsweep
import Aero.Drag.FormAndFrictional
main :: IO ()
main = do
let config :: Config Double
config = gaCruiseConfig
k = formFactorMarkup (cruise_mach config) (bodyFineness config)
cF = cF_skinFriction config
cD_formAndFrictional' = cD_formAndFrictional config
sWet = wettedArea config
sWing = exposedWingArea_ft2 config
putStrLn "-------------------------- configuration: -----------------------------"
print config
putStrLn "\n------------------------- wetted area: ------------------------------"
printWettedArea config
putStrLn "\n------------------------ upsweep drag: ------------------------------"
let cD_pUpsweep' = cD_pUpsweep config
cD_upsweep' = cD_upsweep config
_ <- cD_pUpsweep' `seq` printf "Cd upsweep referenced to fuselage: %.7f\n" cD_pUpsweep'
_ <- cD_upsweep' `seq` printf "Cd upsweep referenced to wing: %.7f\n" cD_upsweep'
putStrLn "\n---------------------- frictional drag: -----------------------------"
_ <- cF `seq` printf "skin friction coeff Cf referenced to wetted area:\t%.5f\n" cF
_ <- sWet `seq` printf "skin friction coeff Cf referenced to wing:\t\t%.5f\n\n" (cF*sWet/sWing)
putStrLn "\n------------------------- form drag: --------------------------------"
_ <- k `seq` printf "form factor k:\t\t\t\t%.5f\n" k
_ <- printf "Cd_form referenced to wetted area:\t%.5f\n" ((k-1)*cF)
_ <- printf "Cd_form referenced to wing area:\t%.5f\n" ((k-1)*cF*sWet/sWing)
putStrLn "\n----------------- frictional drag + form drag: ----------------------"
_ <- cD_formAndFrictional' `seq` printf "Cd_frictional_form referenced to wing:\t%.5f\n" cD_formAndFrictional'
putStrLn "\n---- total parasitic drag (for now: frictional + form + upsweep): ----"
_ <- printf "Cd_parasitic referenced to wing:\t%.5f\n\n" (cD_formAndFrictional' + cD_upsweep')
-- make some plots
--plotCfModel
--plotFormFactorModel
return ()
plotCfModel :: IO ()
plotCfModel = do
let line = plot_lines_values ^= [[ (LogValue re, LogValue (cfOfReynolds' re))
| y <- [5,5.1..9::Double], let re = 10**y]]
$ plot_lines_title ^= "cf"
$ defaultPlotLines
chart = layout1_title ^= "cf vs Reynolds"
$ layout1_plots ^= [Left (toPlot line)]
$ defaultLayout1
renderableToWindow (toRenderable chart) 640 480
_ <- renderableToPNGFile (toRenderable chart) 640 480 "cf_model.png"
return ()
plotFormFactorModel :: IO ()
plotFormFactorModel = do
let line = plot_lines_values ^= [[ (fr, formFactorMarkup 0.55 fr) | fr <- [4,4.1..10::Double]]]
$ plot_lines_title ^= "k"
$ defaultPlotLines
chart = layout1_title ^= "form factor markup k vs body fineness ratio (mach 0.55)"
$ layout1_plots ^= [Left (toPlot line)]
$ defaultLayout1
renderableToWindow (toRenderable chart) 640 480
_ <- renderableToPNGFile (toRenderable chart) 640 480 "k_model.png"
return ()
| ghorn/conceptual-design | ParasiticSummary.hs | bsd-3-clause | 3,282 | 0 | 20 | 657 | 762 | 382 | 380 | 64 | 1 |
-- Copyright (c) 2016-present, Facebook, Inc.
-- All rights reserved.
--
-- This source code is licensed under the BSD-style license found in the
-- LICENSE file in the root directory of this source tree.
{-# LANGUAGE OverloadedStrings #-}
module Duckling.Quantity.MN.Corpus
( corpus
) where
import Data.String
import Prelude
import Duckling.Locale
import Duckling.Resolve
import Duckling.Quantity.Types
import Duckling.Testing.Types
corpus :: Corpus
corpus = (testContext {locale = makeLocale MN Nothing}, testOptions, allExamples)
allExamples :: [Example]
allExamples = concat
[ examples (simple Pound 2 Nothing)
[ "2 фунт"
]
, examples (simple Gram 2 Nothing)
[ "2 грамм"
, "хоёр грамм"
, "2000 миллиграмм"
, "2000 мг"
]
, examples (simple Gram 1000 Nothing)
[ "килограмм"
, "кг"
]
, examples (simple Gram 2000 Nothing)
[ "2 килограмм"
, "2 кг"
, "2000 грамм"
]
, examples (simple Pound 1 Nothing)
[ "фунт"
, "1 фунт"
]
, examples (simple Ounce 2 Nothing)
[ "2 унц"
]
, examples (simple Gram 500 Nothing)
[ "500 грамм"
, "500г"
, "500 г"
, "0.5 кг"
]
]
| facebookincubator/duckling | Duckling/Quantity/MN/Corpus.hs | bsd-3-clause | 1,472 | 0 | 9 | 512 | 274 | 159 | 115 | 37 | 1 |
import Control.Applicative
data List a = Empty | Cons a (List a)
instance Functor List where
-- fmap :: (a -> b) -> [a] -> [b]
fmap f Empty = Empty
fmap f (Cons a b) = Cons (f a) (fmap f b)
instance Applicative List where
--pure :: a -> [a]
pure a = Cons a Empty
-- (<*>) :: [a -> b] -> [a] -> [b]
_ <*> Empty = Empty
Empty <*> _ = Empty
(Cons a Empty) <*> (Cons b Empty) = Cons (a b) Empty
(Cons a t) <*> (Cons b t2) = Cons (a b) (t <*> t2)
instance Alternative List where
empty = Empty
Empty <|> _ = Empty
(Cons a t) <|> b = (Cons a (t <|> b))
join :: List (List a) -> List a
join Empty = Empty
join (Cons a b) = a <|> join b
instance Monad List where
return = pure
a >>= f = join (fmap f a)
| gahara/parsers-for-dummies | test2.hs | bsd-3-clause | 725 | 1 | 9 | 194 | 356 | 176 | 180 | -1 | -1 |
{-|
Module: FRP.Timeless.Framework.UI.Scene
Copyright: (c) 2015 Rongcui Dong
License: BSD3
Maintainer: Rongcui Dong <[email protected]>
-}
module FRP.Timeless.Framework.UI.Scene
where
import FRP.Timeless
import FRP.Timeless.Framework.UI.Events
newtype Scene = Scene {
sceneBox :: forall m s. Monad m => Signal s m UIInput ()
}
| carldong/timeless-SDL | src/FRP/Timeless/Framework/UI/Scene.hs | bsd-3-clause | 351 | 0 | 10 | 65 | 60 | 38 | 22 | -1 | -1 |
module Main where
import Ivory.Tower.Config
import Ivory.OS.FreeRTOS.Tower.STM32
import LDrive.Platforms
import LDrive.Tests.ADCMulti (app)
main :: IO ()
main = compileTowerSTM32FreeRTOS testplatform_stm32 p $
app (stm32config_clock . testplatform_stm32)
testplatform_adcs
testplatform_pwm
testplatform_uart
testplatform_leds
where
p topts = getConfig topts testPlatformParser
| sorki/odrive | test/ADCMultiTest.hs | bsd-3-clause | 441 | 0 | 8 | 101 | 91 | 51 | 40 | 13 | 1 |
{-# LANGUAGE ScopedTypeVariables #-}
module Network.HaskellNet.SMTP
( -- * Types
Command(..)
, Response(..)
, SMTPConnection
-- * Establishing Connection
, connectSMTPPort
, connectSMTP
, connectStream
-- * Operation to a Connection
, sendCommand
, closeSMTP
-- * Other Useful Operations
, sendMail
, doSMTPPort
, doSMTP
, doSMTPStream
, sendMimeMail
)
where
import Network.HaskellNet.BSStream
import Data.ByteString (ByteString)
import qualified Data.ByteString.Char8 as BS
import Network.BSD (getHostName)
import Network
import Control.Applicative ((<$>))
import Control.Exception
import Control.Monad (unless)
import Data.Char (isDigit)
import Network.HaskellNet.Auth
import System.IO
import Network.Mail.Mime
import qualified Data.ByteString.Lazy as B
import qualified Data.ByteString as S
import qualified Data.Text.Lazy as LT
import qualified Data.Text as T
import Prelude hiding (catch)
data SMTPConnection = SMTPC !BSStream ![ByteString]
data Command = HELO String
| EHLO String
| MAIL String
| RCPT String
| DATA ByteString
| EXPN String
| VRFY String
| HELP String
| AUTH AuthType UserName Password
| NOOP
| RSET
| QUIT
deriving (Show, Eq)
type ReplyCode = Int
data Response = Ok
| SystemStatus
| HelpMessage
| ServiceReady
| ServiceClosing
| UserNotLocal
| CannotVerify
| StartMailInput
| ServiceNotAvailable
| MailboxUnavailable
| ErrorInProcessing
| InsufficientSystemStorage
| SyntaxError
| ParameterError
| CommandNotImplemented
| BadSequence
| ParameterNotImplemented
| MailboxUnavailableError
| UserNotLocalError
| ExceededStorage
| MailboxNotAllowed
| TransactionFailed
deriving (Show, Eq)
-- | connecting SMTP server with the specified name and port number.
connectSMTPPort :: String -- ^ name of the server
-> PortNumber -- ^ port number
-> IO SMTPConnection
connectSMTPPort hostname port =
(handleToStream <$> connectTo hostname (PortNumber port))
>>= connectStream
-- | connecting SMTP server with the specified name and port 25.
connectSMTP :: String -- ^ name of the server
-> IO SMTPConnection
connectSMTP = flip connectSMTPPort 25
tryCommand :: BSStream -> Command -> Int -> ReplyCode
-> IO ByteString
tryCommand st cmd tries expectedReply | tries <= 0 = do
bsClose st
fail $ "cannot execute command " ++ show cmd ++
", expected reply code " ++ show expectedReply
tryCommand st cmd tries expectedReply = do
(code, msg) <- sendCommand (SMTPC st []) cmd
if code == expectedReply then
return msg else
tryCommand st cmd (tries - 1) expectedReply
-- | create SMTPConnection from already connected Stream
connectStream :: BSStream -> IO SMTPConnection
connectStream st =
do (code1, _) <- parseResponse st
unless (code1 == 220) $
do bsClose st
fail "cannot connect to the server"
senderHost <- getHostName
msg <- tryCommand st (EHLO senderHost) 3 250
return (SMTPC st (tail $ BS.lines msg))
parseResponse :: BSStream -> IO (ReplyCode, ByteString)
parseResponse st =
do (code, bdy) <- readLines
return (read $ BS.unpack code, BS.unlines bdy)
where readLines =
do l <- bsGetLine st
let (c, bdy) = BS.span isDigit l
if not (BS.null bdy) && BS.head bdy == '-'
then do (c2, ls) <- readLines
return (c2, (BS.tail bdy:ls))
else return (c, [BS.tail bdy])
-- | send a method to a server
sendCommand :: SMTPConnection -> Command -> IO (ReplyCode, ByteString)
sendCommand (SMTPC conn _) (DATA dat) =
do bsPutCrLf conn $ BS.pack "DATA"
(code, _) <- parseResponse conn
unless (code == 354) $ fail "this server cannot accept any data."
mapM_ sendLine $ BS.lines dat ++ [BS.pack "."]
parseResponse conn
where sendLine l = bsPutCrLf conn l
sendCommand (SMTPC conn _) (AUTH LOGIN username password) =
do bsPutCrLf conn command
(_, _) <- parseResponse conn
bsPutCrLf conn $ BS.pack userB64
(_, _) <- parseResponse conn
bsPutCrLf conn $ BS.pack passB64
parseResponse conn
where command = BS.pack $ "AUTH LOGIN"
(userB64, passB64) = login username password
sendCommand (SMTPC conn _) (AUTH at username password) =
do bsPutCrLf conn command
(code, msg) <- parseResponse conn
unless (code == 334) $ fail "authentication failed."
bsPutCrLf conn $ BS.pack $ auth at (BS.unpack msg) username password
parseResponse conn
where command = BS.pack $ unwords ["AUTH", show at]
sendCommand (SMTPC conn _) meth =
do bsPutCrLf conn $ BS.pack command
parseResponse conn
where command = case meth of
(HELO param) -> "HELO " ++ param
(EHLO param) -> "EHLO " ++ param
(MAIL param) -> "MAIL FROM:<" ++ param ++ ">"
(RCPT param) -> "RCPT TO:<" ++ param ++ ">"
(EXPN param) -> "EXPN " ++ param
(VRFY param) -> "VRFY " ++ param
(HELP msg) -> if null msg
then "HELP\r\n"
else "HELP " ++ msg
NOOP -> "NOOP"
RSET -> "RSET"
QUIT -> "QUIT"
(DATA _) ->
error "BUG: DATA pattern should be matched by sendCommand patterns"
(AUTH _ _ _) ->
error "BUG: AUTH pattern should be matched by sendCommand patterns"
-- | close the connection. This function send the QUIT method, so you
-- do not have to QUIT method explicitly.
closeSMTP :: SMTPConnection -> IO ()
closeSMTP (SMTPC conn _) = bsClose conn
{-
I must be being stupid here
I can't seem to be able to catch the exception arising from the
connection already being closed this would be the correct way to do it
but instead we're being naughty above by just closes the connection
without first sending QUIT
closeSMTP c@(SMTPC conn _) =
do sendCommand c QUIT
bsClose conn `catch` \(_ :: IOException) -> return ()
-}
-- | sending a mail to a server. This is achieved by sendMessage. If
-- something is wrong, it raises an IOexception.
sendMail :: String -- ^ sender mail
-> [String] -- ^ receivers
-> ByteString -- ^ data
-> SMTPConnection
-> IO ()
sendMail sender receivers dat conn =
catcher `handle` mainProc
where mainProc =
do (250, _) <- sendCommand conn (MAIL sender)
vals <- mapM (sendCommand conn . RCPT) receivers
unless (all ((==250) . fst) vals) $ fail "sendMail error"
(250, _) <- sendCommand conn (DATA dat)
return ()
catcher e@(PatternMatchFail _) = throwIO e
-- | doSMTPPort open a connection, and do an IO action with the
-- connection, and then close it.
doSMTPPort :: String -> PortNumber -> (SMTPConnection -> IO a) -> IO a
doSMTPPort host port execution =
bracket (connectSMTPPort host port) closeSMTP execution
-- | doSMTP is similar to doSMTPPort, except that it does not require
-- port number but connects to the server with port 25.
doSMTP :: String -> (SMTPConnection -> IO a) -> IO a
doSMTP host execution = doSMTPPort host 25 execution
-- | doSMTPStream is similar to doSMTPPort, except that its argument
-- is a Stream data instead of hostname and port number.
doSMTPStream :: BSStream -> (SMTPConnection -> IO a) -> IO a
doSMTPStream s execution = bracket (connectStream s) closeSMTP execution
sendMimeMail :: String -> String -> String -> LT.Text
-> LT.Text -> [(T.Text, FilePath)] -> SMTPConnection -> IO ()
sendMimeMail to from subject plainBody htmlBody attachments con = do
myMail <- simpleMail (Address Nothing $ T.pack to) (Address Nothing
$ T.pack from)
(T.pack subject) plainBody htmlBody attachments
renderedMail <- renderMail' myMail
sendMail from [to] (lazyToStrict renderedMail) con
closeSMTP con
-- haskellNet uses strict bytestrings
-- TODO: look at making haskellnet lazy
lazyToStrict :: B.ByteString -> S.ByteString
lazyToStrict = S.concat . B.toChunks
crlf :: BS.ByteString
crlf = BS.pack "\r\n"
bsPutCrLf :: BSStream -> ByteString -> IO ()
bsPutCrLf h s = bsPut h s >> bsPut h crlf >> bsFlush h
| danchoi/imapget | src/Network/HaskellNet/SMTP.hs | bsd-3-clause | 9,059 | 0 | 17 | 2,875 | 2,210 | 1,143 | 1,067 | 197 | 13 |
{-# LANGUAGE ExplicitForAll #-}
-- | This game mode lets you manage your own input. Pressing ESC will not abort the program.
-- You also don't get automatic pan and zoom controls like with `displayInWindow`.
module Graphics.Gloss.Interface.IO.Game
( module Graphics.Gloss.Data.Display
, module Graphics.Gloss.Data.Picture
, module Graphics.Gloss.Data.Color
, playIO
, Event(..), Key(..), SpecialKey(..), MouseButton(..), KeyState(..), Modifiers(..))
where
import Graphics.Gloss.Data.Display
import Graphics.Gloss.Data.Picture
import Graphics.Gloss.Data.Color
import Graphics.Gloss.Internals.Interface.Game
import Graphics.Gloss.Internals.Interface.Backend
-- | Play a game in a window, using IO actions to build the pictures.
playIO :: forall world
. Display -- ^ Display mode.
-> Color -- ^ Background color.
-> Int -- ^ Number of simulation steps to take for each second of real time.
-> world -- ^ The initial world.
-> (world -> IO Picture) -- ^ An action to convert the world a picture.
-> (Event -> world -> IO world) -- ^ A function to handle input events.
-> (Float -> world -> IO world) -- ^ A function to step the world one iteration.
-- It is passed the period of time (in seconds) needing to be advanced.
-> IO ()
playIO display backColor simResolution
worldStart worldToPicture worldHandleEvent worldAdvance
= playWithBackendIO defaultBackendState
display backColor simResolution
worldStart worldToPicture worldHandleEvent worldAdvance
False
| gscalzo/HaskellTheHardWay | gloss-try/gloss-master/gloss/Graphics/Gloss/Interface/IO/Game.hs | mit | 1,743 | 0 | 16 | 491 | 249 | 157 | 92 | 27 | 1 |
-----------------------------------------------------------------------------
-- |
-- Module : Distribution.Simple.Register
-- Copyright : Isaac Jones 2003-2004
--
-- Maintainer : [email protected]
-- Portability : portable
--
-- This module deals with registering and unregistering packages. There are a
-- couple ways it can do this, one is to do it directly. Another is to generate
-- a script that can be run later to do it. The idea here being that the user
-- is shielded from the details of what command to use for package registration
-- for a particular compiler. In practice this aspect was not especially
-- popular so we also provide a way to simply generate the package registration
-- file which then must be manually passed to @ghc-pkg@. It is possible to
-- generate registration information for where the package is to be installed,
-- or alternatively to register the package inplace in the build tree. The
-- latter is occasionally handy, and will become more important when we try to
-- build multi-package systems.
--
-- This module does not delegate anything to the per-compiler modules but just
-- mixes it all in in this module, which is rather unsatisfactory. The script
-- generation and the unregister feature are not well used or tested.
{- Copyright (c) 2003-2004, Isaac Jones
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are
met:
* Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
* Redistributions in binary form must reproduce the above
copyright notice, this list of conditions and the following
disclaimer in the documentation and/or other materials provided
with the distribution.
* Neither the name of Isaac Jones nor the names of other
contributors may be used to endorse or promote products derived
from this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. -}
module Distribution.Simple.Register (
register,
unregister,
initPackageDB,
invokeHcPkg,
registerPackage,
generateRegistrationInfo,
inplaceInstalledPackageInfo,
absoluteInstalledPackageInfo,
generalInstalledPackageInfo,
) where
import Distribution.Simple.LocalBuildInfo
( LocalBuildInfo(..), ComponentLocalBuildInfo(..)
, ComponentName(..), getComponentLocalBuildInfo
, LibraryName(..)
, InstallDirs(..), absoluteInstallDirs )
import Distribution.Simple.BuildPaths (haddockName)
import qualified Distribution.Simple.GHC as GHC
import qualified Distribution.Simple.LHC as LHC
import qualified Distribution.Simple.Hugs as Hugs
import qualified Distribution.Simple.UHC as UHC
import Distribution.Simple.Compiler
( compilerVersion, Compiler, CompilerFlavor(..), compilerFlavor
, PackageDBStack, registrationPackageDB )
import Distribution.Simple.Program
( ProgramConfiguration, ConfiguredProgram
, runProgramInvocation, requireProgram, lookupProgram
, ghcPkgProgram, lhcPkgProgram )
import Distribution.Simple.Program.Script
( invocationAsSystemScript )
import qualified Distribution.Simple.Program.HcPkg as HcPkg
import Distribution.Simple.Setup
( RegisterFlags(..), CopyDest(..)
, fromFlag, fromFlagOrDefault, flagToMaybe )
import Distribution.PackageDescription
( PackageDescription(..), Library(..), BuildInfo(..), hcOptions )
import Distribution.Package
( Package(..), packageName, InstalledPackageId(..) )
import Distribution.InstalledPackageInfo
( InstalledPackageInfo, InstalledPackageInfo_(InstalledPackageInfo)
, showInstalledPackageInfo )
import qualified Distribution.InstalledPackageInfo as IPI
import Distribution.Simple.Utils
( writeUTF8File, writeFileAtomic, setFileExecutable
, die, notice, setupMessage )
import Distribution.System
( OS(..), buildOS )
import Distribution.Text
( display )
import Distribution.Version ( Version(..) )
import Distribution.Verbosity as Verbosity
( Verbosity, normal )
import Distribution.Compat.Exception
( tryIO )
import System.FilePath ((</>), (<.>), isAbsolute)
import System.Directory
( getCurrentDirectory, removeDirectoryRecursive )
import Data.Maybe
( isJust, fromMaybe, maybeToList )
import Data.List
( partition, nub )
import qualified Data.ByteString.Lazy.Char8 as BS.Char8
-- -----------------------------------------------------------------------------
-- Registration
register :: PackageDescription -> LocalBuildInfo
-> RegisterFlags -- ^Install in the user's database?; verbose
-> IO ()
register pkg@PackageDescription { library = Just lib } lbi regFlags
= do
let clbi = getComponentLocalBuildInfo lbi CLibName
installedPkgInfo <- generateRegistrationInfo
verbosity pkg lib lbi clbi inplace distPref
-- Three different modes:
case () of
_ | modeGenerateRegFile -> writeRegistrationFile installedPkgInfo
| modeGenerateRegScript -> writeRegisterScript installedPkgInfo
| otherwise -> registerPackage verbosity
installedPkgInfo pkg lbi inplace packageDbs
where
modeGenerateRegFile = isJust (flagToMaybe (regGenPkgConf regFlags))
regFile = fromMaybe (display (packageId pkg) <.> "conf")
(fromFlag (regGenPkgConf regFlags))
modeGenerateRegScript = fromFlag (regGenScript regFlags)
inplace = fromFlag (regInPlace regFlags)
-- FIXME: there's really no guarantee this will work.
-- registering into a totally different db stack can
-- fail if dependencies cannot be satisfied.
packageDbs = nub $ withPackageDB lbi
++ maybeToList (flagToMaybe (regPackageDB regFlags))
distPref = fromFlag (regDistPref regFlags)
verbosity = fromFlag (regVerbosity regFlags)
writeRegistrationFile installedPkgInfo = do
notice verbosity ("Creating package registration file: " ++ regFile)
writeUTF8File regFile (showInstalledPackageInfo installedPkgInfo)
writeRegisterScript installedPkgInfo =
case compilerFlavor (compiler lbi) of
GHC -> do (ghcPkg, _) <- requireProgram verbosity ghcPkgProgram (withPrograms lbi)
writeHcPkgRegisterScript verbosity installedPkgInfo ghcPkg packageDbs
LHC -> do (lhcPkg, _) <- requireProgram verbosity lhcPkgProgram (withPrograms lbi)
writeHcPkgRegisterScript verbosity installedPkgInfo lhcPkg packageDbs
Hugs -> notice verbosity "Registration scripts not needed for hugs"
JHC -> notice verbosity "Registration scripts not needed for jhc"
NHC -> notice verbosity "Registration scripts not needed for nhc98"
UHC -> notice verbosity "Registration scripts not needed for uhc"
_ -> die "Registration scripts are not implemented for this compiler"
register _ _ regFlags = notice verbosity "No package to register"
where
verbosity = fromFlag (regVerbosity regFlags)
generateRegistrationInfo :: Verbosity
-> PackageDescription
-> Library
-> LocalBuildInfo
-> ComponentLocalBuildInfo
-> Bool
-> FilePath
-> IO InstalledPackageInfo
generateRegistrationInfo verbosity pkg lib lbi clbi inplace distPref = do
--TODO: eliminate pwd!
pwd <- getCurrentDirectory
--TODO: the method of setting the InstalledPackageId is compiler specific
-- this aspect should be delegated to a per-compiler helper.
let comp = compiler lbi
ipid <-
case compilerFlavor comp of
GHC | compilerVersion comp >= Version [6,11] [] -> do
s <- GHC.libAbiHash verbosity pkg lbi lib clbi
return (InstalledPackageId (display (packageId pkg) ++ '-':s))
_other -> do
return (InstalledPackageId (display (packageId pkg)))
let installedPkgInfo
| inplace = inplaceInstalledPackageInfo pwd distPref
pkg lib lbi clbi
| otherwise = absoluteInstalledPackageInfo
pkg lib lbi clbi
return installedPkgInfo{ IPI.installedPackageId = ipid }
-- | Create an empty package DB at the specified location.
initPackageDB :: Verbosity -> Compiler -> ProgramConfiguration -> FilePath
-> IO ()
initPackageDB verbosity comp conf dbPath =
case (compilerFlavor comp) of
GHC -> GHC.initPackageDB verbosity conf dbPath
_ -> die "Distribution.Simple.Register.initPackageDB: \
\not implemented for this compiler"
-- | Run @hc-pkg@ using a given package DB stack, directly forwarding the
-- provided command-line arguments to it.
invokeHcPkg :: Verbosity -> Compiler -> ProgramConfiguration -> PackageDBStack
-> [String] -> IO ()
invokeHcPkg verbosity comp conf dbStack extraArgs =
case (compilerFlavor comp) of
GHC -> GHC.invokeHcPkg verbosity conf dbStack extraArgs
_ -> die "Distribution.Simple.Register.invokeHcPkg: \
\not implemented for this compiler"
registerPackage :: Verbosity
-> InstalledPackageInfo
-> PackageDescription
-> LocalBuildInfo
-> Bool
-> PackageDBStack
-> IO ()
registerPackage verbosity installedPkgInfo pkg lbi inplace packageDbs = do
let msg = if inplace
then "In-place registering"
else "Registering"
setupMessage verbosity msg (packageId pkg)
case compilerFlavor (compiler lbi) of
GHC -> GHC.registerPackage verbosity installedPkgInfo pkg lbi inplace packageDbs
LHC -> LHC.registerPackage verbosity installedPkgInfo pkg lbi inplace packageDbs
Hugs -> Hugs.registerPackage verbosity installedPkgInfo pkg lbi inplace packageDbs
UHC -> UHC.registerPackage verbosity installedPkgInfo pkg lbi inplace packageDbs
JHC -> notice verbosity "Registering for jhc (nothing to do)"
NHC -> notice verbosity "Registering for nhc98 (nothing to do)"
_ -> die "Registering is not implemented for this compiler"
writeHcPkgRegisterScript :: Verbosity
-> InstalledPackageInfo
-> ConfiguredProgram
-> PackageDBStack
-> IO ()
writeHcPkgRegisterScript verbosity installedPkgInfo hcPkg packageDbs = do
let invocation = HcPkg.reregisterInvocation hcPkg Verbosity.normal
packageDbs (Right installedPkgInfo)
regScript = invocationAsSystemScript buildOS invocation
notice verbosity ("Creating package registration script: " ++ regScriptFileName)
writeUTF8File regScriptFileName regScript
setFileExecutable regScriptFileName
regScriptFileName :: FilePath
regScriptFileName = case buildOS of
Windows -> "register.bat"
_ -> "register.sh"
-- -----------------------------------------------------------------------------
-- Making the InstalledPackageInfo
-- | Construct 'InstalledPackageInfo' for a library in a package, given a set
-- of installation directories.
--
generalInstalledPackageInfo
:: ([FilePath] -> [FilePath]) -- ^ Translate relative include dir paths to
-- absolute paths.
-> PackageDescription
-> Library
-> ComponentLocalBuildInfo
-> InstallDirs FilePath
-> InstalledPackageInfo
generalInstalledPackageInfo adjustRelIncDirs pkg lib clbi installDirs =
InstalledPackageInfo {
--TODO: do not open-code this conversion from PackageId to InstalledPackageId
IPI.installedPackageId = InstalledPackageId (display (packageId pkg)),
IPI.sourcePackageId = packageId pkg,
IPI.license = license pkg,
IPI.copyright = copyright pkg,
IPI.maintainer = maintainer pkg,
IPI.author = author pkg,
IPI.stability = stability pkg,
IPI.homepage = homepage pkg,
IPI.pkgUrl = pkgUrl pkg,
IPI.synopsis = synopsis pkg,
IPI.description = description pkg,
IPI.category = category pkg,
IPI.exposed = libExposed lib,
IPI.exposedModules = exposedModules lib,
IPI.hiddenModules = otherModules bi,
IPI.trusted = IPI.trusted IPI.emptyInstalledPackageInfo,
IPI.importDirs = [ libdir installDirs | hasModules ],
IPI.libraryDirs = if hasLibrary
then libdir installDirs : extraLibDirs bi
else extraLibDirs bi,
IPI.hsLibraries = [ libname
| LibraryName libname <- componentLibraries clbi
, hasLibrary ],
IPI.extraLibraries = extraLibs bi,
IPI.extraGHCiLibraries = [],
IPI.includeDirs = absinc ++ adjustRelIncDirs relinc,
IPI.includes = includes bi,
IPI.depends = map fst (componentPackageDeps clbi),
IPI.hugsOptions = hcOptions Hugs bi,
IPI.ccOptions = [], -- Note. NOT ccOptions bi!
-- We don't want cc-options to be propagated
-- to C compilations in other packages.
IPI.ldOptions = ldOptions bi,
IPI.frameworkDirs = [],
IPI.frameworks = frameworks bi,
IPI.haddockInterfaces = [haddockdir installDirs </> haddockName pkg],
IPI.haddockHTMLs = [htmldir installDirs]
}
where
bi = libBuildInfo lib
(absinc, relinc) = partition isAbsolute (includeDirs bi)
hasModules = not $ null (exposedModules lib)
&& null (otherModules bi)
hasLibrary = hasModules || not (null (cSources bi))
-- | Construct 'InstalledPackageInfo' for a library that is inplace in the
-- build tree.
--
-- This function knows about the layout of inplace packages.
--
inplaceInstalledPackageInfo :: FilePath -- ^ top of the build tree
-> FilePath -- ^ location of the dist tree
-> PackageDescription
-> Library
-> LocalBuildInfo
-> ComponentLocalBuildInfo
-> InstalledPackageInfo
inplaceInstalledPackageInfo inplaceDir distPref pkg lib lbi clbi =
generalInstalledPackageInfo adjustRelativeIncludeDirs pkg lib clbi
installDirs
where
adjustRelativeIncludeDirs = map (inplaceDir </>)
installDirs =
(absoluteInstallDirs pkg lbi NoCopyDest) {
libdir = inplaceDir </> buildDir lbi,
datadir = inplaceDir,
datasubdir = distPref,
docdir = inplaceDocdir,
htmldir = inplaceHtmldir,
haddockdir = inplaceHtmldir
}
inplaceDocdir = inplaceDir </> distPref </> "doc"
inplaceHtmldir = inplaceDocdir </> "html" </> display (packageName pkg)
-- | Construct 'InstalledPackageInfo' for the final install location of a
-- library package.
--
-- This function knows about the layout of installed packages.
--
absoluteInstalledPackageInfo :: PackageDescription
-> Library
-> LocalBuildInfo
-> ComponentLocalBuildInfo
-> InstalledPackageInfo
absoluteInstalledPackageInfo pkg lib lbi clbi =
generalInstalledPackageInfo adjustReativeIncludeDirs pkg lib clbi installDirs
where
-- For installed packages we install all include files into one dir,
-- whereas in the build tree they may live in multiple local dirs.
adjustReativeIncludeDirs _
| null (installIncludes bi) = []
| otherwise = [includedir installDirs]
bi = libBuildInfo lib
installDirs = absoluteInstallDirs pkg lbi NoCopyDest
-- -----------------------------------------------------------------------------
-- Unregistration
unregister :: PackageDescription -> LocalBuildInfo -> RegisterFlags -> IO ()
unregister pkg lbi regFlags = do
let pkgid = packageId pkg
genScript = fromFlag (regGenScript regFlags)
verbosity = fromFlag (regVerbosity regFlags)
packageDb = fromFlagOrDefault (registrationPackageDB (withPackageDB lbi))
(regPackageDB regFlags)
installDirs = absoluteInstallDirs pkg lbi NoCopyDest
setupMessage verbosity "Unregistering" pkgid
case compilerFlavor (compiler lbi) of
GHC ->
let Just ghcPkg = lookupProgram ghcPkgProgram (withPrograms lbi)
invocation = HcPkg.unregisterInvocation ghcPkg Verbosity.normal
packageDb pkgid
in if genScript
then writeFileAtomic unregScriptFileName
(BS.Char8.pack $ invocationAsSystemScript buildOS invocation)
else runProgramInvocation verbosity invocation
Hugs -> do
_ <- tryIO $ removeDirectoryRecursive (libdir installDirs)
return ()
NHC -> do
_ <- tryIO $ removeDirectoryRecursive (libdir installDirs)
return ()
_ ->
die ("only unregistering with GHC and Hugs is implemented")
unregScriptFileName :: FilePath
unregScriptFileName = case buildOS of
Windows -> "unregister.bat"
_ -> "unregister.sh"
| jwiegley/ghc-release | libraries/Cabal/cabal/Distribution/Simple/Register.hs | gpl-3.0 | 18,469 | 0 | 22 | 4,859 | 3,029 | 1,620 | 1,409 | 283 | 8 |
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE helpset PUBLIC "-//Sun Microsystems Inc.//DTD JavaHelp HelpSet Version 2.0//EN" "http://java.sun.com/products/javahelp/helpset_2_0.dtd">
<helpset version="2.0" xml:lang="tr-TR">
<title>Aktif Tarama Kuralları - Blpha | ZAP Uzantısı</title>
<maps>
<homeID>top</homeID>
<mapref location="map.jhm"/>
</maps>
<view>
<name>TOC</name>
<label>İçindekiler</label>
<type>org.zaproxy.zap.extension.help.ZapTocView</type>
<data>toc.xml</data>
</view>
<view>
<name>Index</name>
<label>İçerik</label>
<type>javax.help.IndexView</type>
<data>index.xml</data>
</view>
<view>
<name>Search</name>
<label>Arama</label>
<type>javax.help.SearchView</type>
<data engine="com.sun.java.help.search.DefaultSearchEngine">
JavaHelpSearch
</data>
</view>
<view>
<name>Favorites</name>
<label>Favoriler</label>
<type>javax.help.FavoritesView</type>
</view>
</helpset> | veggiespam/zap-extensions | addOns/simpleexample/src/main/javahelp/org/zaproxy/zap/extension/simpleexample/resources/help_tr_TR/helpset_tr_TR.hs | apache-2.0 | 1,001 | 80 | 67 | 163 | 434 | 218 | 216 | -1 | -1 |
{-# OPTIONS_HADDOCK hide #-}
{-# LANGUAGE CPP #-}
#include "fusion-phases.h"
-- | PR instance for tuples.
module Data.Array.Parallel.PArray.PData.Tuple4
( PData(..), PDatas(..)
, zip4PD)
where
import Data.Array.Parallel.Pretty
import Data.Array.Parallel.PArray.PData.Base
import Data.Array.Parallel.PArray.PData.Nested
import GHC.Exts
import Prelude hiding (zip, unzip)
import qualified Data.Typeable as T
import qualified Data.Vector as V
import qualified Data.List as P
-------------------------------------------------------------------------------
data instance PData (a, b, c, d)
= PTuple4 (PData a) (PData b) (PData c) (PData d)
data instance PDatas (a, b, c, d)
= PTuple4s (PDatas a) (PDatas b) (PDatas c) (PDatas d)
-- PR -------------------------------------------------------------------------
instance (PR a, PR b, PR c, PR d) => PR (a, b, c, d) where
{-# NOINLINE validPR #-}
validPR (PTuple4 xs ys zs ds)
= validPR xs && validPR ys && validPR zs && validPR ds
{-# NOINLINE nfPR #-}
nfPR (PTuple4 arr1 arr2 arr3 arr4)
= nfPR arr1 `seq` nfPR arr2 `seq` nfPR arr3 `seq` nfPR arr4 `seq` ()
{-# NOINLINE similarPR #-}
similarPR (x1, y1, z1, d1) (x2, y2, z2, d2)
= similarPR x1 x2
&& similarPR y1 y2
&& similarPR z1 z2
&& similarPR d1 d2
{-# NOINLINE coversPR #-}
coversPR weak (PTuple4 arr1 arr2 arr3 arr4) ix
= coversPR weak arr1 ix
&& coversPR weak arr2 ix
&& coversPR weak arr3 ix
&& coversPR weak arr4 ix
{-# NOINLINE pprpPR #-}
pprpPR (x, y, z, d)
= text "Tuple4 "
<> vcat [ pprpPR x
, pprpPR y
, pprpPR z
, pprpPR d ]
{-# NOINLINE pprpDataPR #-}
pprpDataPR (PTuple4 xs ys zs ds)
= text "PTuple4 "
<> vcat [ pprpDataPR xs
, pprpDataPR ys
, pprpDataPR zs
, pprpDataPR ds]
{-# NOINLINE typeRepPR #-}
typeRepPR x@(a, b, c, d)
= T.typeOf4 x
`T.mkAppTy` (typeRepPR a)
`T.mkAppTy` (typeRepPR b)
`T.mkAppTy` (typeRepPR c)
`T.mkAppTy` (typeRepPR d)
{-# NOINLINE typeRepDataPR #-}
typeRepDataPR (PTuple4 as bs cs ds)
= T.typeOf4 ((), (), (), ())
`T.mkAppTy` (typeRepDataPR as)
`T.mkAppTy` (typeRepDataPR bs)
`T.mkAppTy` (typeRepDataPR cs)
`T.mkAppTy` (typeRepDataPR ds)
{-# NOINLINE typeRepDatasPR #-}
typeRepDatasPR (PTuple4s as bs cs ds)
= T.typeOf4 ((), (), (), ())
`T.mkAppTy` (typeRepDatasPR as)
`T.mkAppTy` (typeRepDatasPR bs)
`T.mkAppTy` (typeRepDatasPR cs)
`T.mkAppTy` (typeRepDatasPR ds)
-- Constructors -------------------------------
{-# INLINE_PDATA emptyPR #-}
emptyPR
= PTuple4 emptyPR emptyPR emptyPR emptyPR
{-# INLINE_PDATA replicatePR #-}
replicatePR len (x, y, z, d)
= PTuple4 (replicatePR len x)
(replicatePR len y)
(replicatePR len z)
(replicatePR len d)
{-# INLINE_PDATA replicatesPR #-}
replicatesPR lens (PTuple4 arr1 arr2 arr3 arr4)
= PTuple4 (replicatesPR lens arr1)
(replicatesPR lens arr2)
(replicatesPR lens arr3)
(replicatesPR lens arr4)
{-# INLINE_PDATA appendPR #-}
appendPR (PTuple4 arr11 arr12 arr13 arr14)
(PTuple4 arr21 arr22 arr23 arr24)
= PTuple4 (arr11 `appendPR` arr21)
(arr12 `appendPR` arr22)
(arr13 `appendPR` arr23)
(arr14 `appendPR` arr24)
{-# INLINE_PDATA appendvsPR #-}
appendvsPR segdResult segd1 (PTuple4s arrs11 arrs12 arrs13 arrs14)
segd2 (PTuple4s arrs21 arrs22 arrs23 arrs24)
= PTuple4 (appendvsPR segdResult segd1 arrs11 segd2 arrs21)
(appendvsPR segdResult segd1 arrs12 segd2 arrs22)
(appendvsPR segdResult segd1 arrs13 segd2 arrs23)
(appendvsPR segdResult segd1 arrs14 segd2 arrs24)
-- Projections ---------------------------------
{-# INLINE_PDATA lengthPR #-}
lengthPR (PTuple4 arr1 _ _ _)
= lengthPR arr1
{-# INLINE_PDATA indexPR #-}
indexPR (PTuple4 arr1 arr2 arr3 arr4) ix
= ( indexPR arr1 ix
, indexPR arr2 ix
, indexPR arr3 ix
, indexPR arr4 ix)
{-# INLINE_PDATA indexsPR #-}
indexsPR (PTuple4s xs ys zs ds) srcixs
= PTuple4 (indexsPR xs srcixs)
(indexsPR ys srcixs)
(indexsPR zs srcixs)
(indexsPR ds srcixs)
{-# INLINE_PDATA indexvsPR #-}
indexvsPR (PTuple4s xs ys zs ds) vsegd srcixs
= PTuple4 (indexvsPR xs vsegd srcixs)
(indexvsPR ys vsegd srcixs)
(indexvsPR zs vsegd srcixs)
(indexvsPR ds vsegd srcixs)
{-# INLINE_PDATA extractPR #-}
extractPR (PTuple4 arr1 arr2 arr3 arr4) start len
= PTuple4 (extractPR arr1 start len)
(extractPR arr2 start len)
(extractPR arr3 start len)
(extractPR arr4 start len)
{-# INLINE_PDATA extractssPR #-}
extractssPR (PTuple4s xs ys zs ds) ussegd
= PTuple4 (extractssPR xs ussegd)
(extractssPR ys ussegd)
(extractssPR zs ussegd)
(extractssPR ds ussegd)
{-# INLINE_PDATA extractvsPR #-}
extractvsPR (PTuple4s xs ys zs ds) uvsegd
= PTuple4 (extractvsPR xs uvsegd)
(extractvsPR ys uvsegd)
(extractvsPR zs uvsegd)
(extractvsPR ds uvsegd)
-- Pack and Combine ---------------------------
{-# INLINE_PDATA packByTagPR #-}
packByTagPR (PTuple4 arr1 arr2 arr3 arr4) tags tag
= PTuple4 (packByTagPR arr1 tags tag)
(packByTagPR arr2 tags tag)
(packByTagPR arr3 tags tag)
(packByTagPR arr4 tags tag)
{-# INLINE_PDATA combine2PR #-}
combine2PR sel (PTuple4 xs1 ys1 zs1 ds1) (PTuple4 xs2 ys2 zs2 ds2)
= PTuple4 (combine2PR sel xs1 xs2)
(combine2PR sel ys1 ys2)
(combine2PR sel zs1 zs2)
(combine2PR sel ds1 ds2)
-- Conversions --------------------------------
{-# NOINLINE fromVectorPR #-}
fromVectorPR vec
= let (xs, ys, zs, ds) = V.unzip4 vec
in PTuple4 (fromVectorPR xs)
(fromVectorPR ys)
(fromVectorPR zs)
(fromVectorPR ds)
{-# NOINLINE toVectorPR #-}
toVectorPR (PTuple4 xs ys zs ds)
= V.zip4 (toVectorPR xs)
(toVectorPR ys)
(toVectorPR zs)
(toVectorPR ds)
-- PData --------------------------------------
{-# INLINE_PDATA emptydPR #-}
emptydPR
= PTuple4s emptydPR
emptydPR
emptydPR
emptydPR
{-# INLINE_PDATA singletondPR #-}
singletondPR (PTuple4 x y z d)
= PTuple4s (singletondPR x)
(singletondPR y)
(singletondPR z)
(singletondPR d)
{-# INLINE_PDATA lengthdPR #-}
lengthdPR (PTuple4s xs _ _ _)
= lengthdPR xs
{-# INLINE_PDATA indexdPR #-}
indexdPR (PTuple4s xs ys zs ds) i
= PTuple4 (indexdPR xs i)
(indexdPR ys i)
(indexdPR zs i)
(indexdPR ds i)
{-# INLINE_PDATA appenddPR #-}
appenddPR (PTuple4s xs1 ys1 zs1 ds1) (PTuple4s xs2 ys2 zs2 ds2)
= PTuple4s (appenddPR xs1 xs2)
(appenddPR ys1 ys2)
(appenddPR zs1 zs2)
(appenddPR ds1 ds2)
{-# NOINLINE fromVectordPR #-}
fromVectordPR vec
= let (xss, yss, zss, dss) = V.unzip4 $ V.map (\(PTuple4 xs ys zs ds) -> (xs, ys, zs, ds)) vec
in PTuple4s (fromVectordPR xss)
(fromVectordPR yss)
(fromVectordPR zss)
(fromVectordPR dss)
{-# NOINLINE toVectordPR #-}
toVectordPR (PTuple4s pdatas1 pdatas2 pdatas3 pdatas4)
= V.zipWith4 PTuple4
(toVectordPR pdatas1)
(toVectordPR pdatas2)
(toVectordPR pdatas3)
(toVectordPR pdatas4)
-- PD Functions ---------------------------------------------------------------
-- | O(1). Zip a pair of arrays into an array of pairs.
zip4PD :: PData a -> PData b -> PData c -> PData d -> PData (a, b, c, d)
zip4PD = PTuple4
{-# INLINE_PA zip4PD #-}
-- Show -----------------------------------------------------------------------
deriving instance (Show (PData a), Show (PData b), Show (PData c), Show (PData d))
=> Show (PData (a, b, c, d))
deriving instance (Show (PDatas a), Show (PDatas b), Show (PDatas c), Show (PDatas d))
=> Show (PDatas (a, b, c, d))
instance ( PR a, PR b, PR c, PR d, Show a, Show b, Show c, Show d
, PprVirtual (PData a), PprVirtual (PData b), PprVirtual (PData c), PprVirtual (PData d))
=> PprVirtual (PData (a, b, c, d)) where
pprv (PTuple4 xs ys zs ds)
= text $ show
$ P.zip4 (V.toList $ toVectorPR xs)
(V.toList $ toVectorPR ys)
(V.toList $ toVectorPR zs)
(V.toList $ toVectorPR ds)
| mainland/dph | dph-lifted-vseg/Data/Array/Parallel/PArray/PData/Tuple4.hs | bsd-3-clause | 9,606 | 0 | 15 | 3,330 | 2,760 | 1,458 | 1,302 | -1 | -1 |
-------------------------------------------------------------------------
--
-- Haskell: The Craft of Functional Programming, 3e
-- Simon Thompson
-- (c) Addison-Wesley, 1996-2011.
--
-- Case study: Parsing expressions
--
-- Note that this is not a monadic approach to parsing.
--
---------------------------------------------------------------------------
module ParsingBasics where
import Data.Char
infixr 5 >*>
--
-- Syntactic types
--
type Var = Char
data Expr = Lit Int | Var Var | Op Op Expr Expr
data Op = Add | Sub | Mul | Div | Mod
--
-- The type of parsers.
--
type Parse a b = [a] -> [(b,[a])]
--
-- Some basic parsers
--
--
-- Fail on any input.
--
none :: Parse a b
none inp = []
--
-- Succeed, returning the value supplied.
--
succeed :: b -> Parse a b
succeed val inp = [(val,inp)]
--
-- token t recognises t as the first value in the input.
--
token :: Eq a => a -> Parse a a
token t (x:xs)
| t==x = [(t,xs)]
| otherwise = []
token t [] = []
--
-- spot whether an element with a particular property is the
-- first element of input.
--
spot :: (a -> Bool) -> Parse a a
spot p (x:xs)
| p x = [(x,xs)]
| otherwise = []
spot p [] = []
--
-- Examples.
--
bracket = token '('
dig = spot isDigit
--
-- Combining parsers
--
--
-- alt p1 p2 recognises anything recogniseed by p1 or by p2.
--
alt :: Parse a b -> Parse a b -> Parse a b
alt p1 p2 inp = p1 inp ++ p2 inp
exam1 = (bracket `alt` dig) "234"
--
-- Apply one parser then the second to the result(s) of the first.
--
(>*>) :: Parse a b -> Parse a c -> Parse a (b,c)
--
(>*>) p1 p2 inp
= [((y,z),rem2) | (y,rem1) <- p1 inp , (z,rem2) <- p2 rem1 ]
--
-- Transform the results of the parses according to the function.
--
build :: Parse a b -> (b -> c) -> Parse a c
build p f inp = [ (f x,rem) | (x,rem) <- p inp ]
--
-- Recognise a list of objects.
--
--
list :: Parse a b -> Parse a [b]
list p = (succeed []) `alt`
((p >*> list p) `build` convert)
where
convert = uncurry (:)
--
-- From the exercises...
--
neList :: Parse a b -> Parse a [b]
neList = neList -- dummy definition
optional :: Parse a b -> Parse a [b]
optional = optional -- dummy definition
nTimes :: Int -> Parse a b -> Parse a [b]
nTimes = nTimes -- dummy definition
--
-- A parser for expressions
--
--
-- The parser has three components, corresponding to the three
-- clauses in the definition of the syntactic type.
--
parser :: Parse Char Expr
parser = (litParse `alt` varParse) `alt` opExpParse
--
-- Spotting variables.
--
varParse :: Parse Char Expr
varParse = spot isVar `build` Var
isVar :: Char -> Bool
isVar x = ('a' <= x && x <= 'z')
--
-- Parsing (fully bracketed) operator applications.
--
opExpParse
= (token '(' >*>
parser >*>
spot isOp >*>
parser >*>
token ')')
`build` makeExpr
makeExpr (_,(e1,(bop,(e2,_)))) = Op (charToOp bop) e1 e2
isOp :: Char -> Bool
isOp = isOp -- dummy definition
charToOp :: Char -> Op
charToOp = charToOp -- dummy definition
--
-- A number is a list of digits with an optional ~ at the front.
--
litParse
= ((optional (token '~')) >*>
(neList (spot isDigit)))
`build` (charlistToExpr.join)
where
join = uncurry (++)
--
-- From the exercises...
--
charlistToExpr :: [Char] -> Expr
charlistToExpr = charlistToExpr -- dummy definition
--
-- A grammar for unbracketed expressions.
--
-- eXpr ::= Int | Var | (eXpr Op eXpr) |
-- lexpr mop mexpr | mexpr aop eXpr
-- lexpr ::= Int | Var | (eXpr Op eXpr)
-- mexpr ::= Int | Var | (eXpr Op eXpr) | lexpr mop mexpr
-- mop ::= 'a' | '/' | '\%'
-- aop ::= '+' | '-'
--
--
-- The top-level parser
--
topLevel :: Parse a b -> [a] -> b
topLevel p inp
= case results of
[] -> error "parse unsuccessful"
_ -> head results
where
results = [ found | (found,[]) <- p inp ]
--
-- The type of commands.
--
data Command = Eval Expr | Assign Var Expr | Null
commandParse :: Parse Char Command
commandParse = commandParse -- dummy definition
--
-- From the exercises.
--
-- tokenList :: [a] -> Parse a [a]
-- spotWhile :: (a -> Bool) -> Parse a [a]
| c089/haskell-craft3e | ParsingBasics.hs | mit | 4,456 | 8 | 11 | 1,298 | 1,221 | 706 | 515 | 75 | 2 |
{-# LANGUAGE DataKinds #-}
{-# LANGUAGE DeriveGeneric #-}
{-# LANGUAGE FlexibleInstances #-}
{-# LANGUAGE GeneralizedNewtypeDeriving #-}
{-# LANGUAGE LambdaCase #-}
{-# LANGUAGE NoImplicitPrelude #-}
{-# LANGUAGE OverloadedStrings #-}
{-# LANGUAGE RecordWildCards #-}
{-# LANGUAGE TypeFamilies #-}
{-# OPTIONS_GHC -fno-warn-unused-imports #-}
-- Module : Network.AWS.SNS.AddPermission
-- Copyright : (c) 2013-2014 Brendan Hay <[email protected]>
-- License : This Source Code Form is subject to the terms of
-- the Mozilla Public License, v. 2.0.
-- A copy of the MPL can be found in the LICENSE file or
-- you can obtain it at http://mozilla.org/MPL/2.0/.
-- Maintainer : Brendan Hay <[email protected]>
-- Stability : experimental
-- Portability : non-portable (GHC extensions)
--
-- Derived from AWS service descriptions, licensed under Apache 2.0.
-- | Adds a statement to a topic's access control policy, granting access for the
-- specified AWS accounts to the specified actions.
--
-- <http://docs.aws.amazon.com/sns/latest/api/API_AddPermission.html>
module Network.AWS.SNS.AddPermission
(
-- * Request
AddPermission
-- ** Request constructor
, addPermission
-- ** Request lenses
, apAWSAccountId
, apActionName
, apLabel
, apTopicArn
-- * Response
, AddPermissionResponse
-- ** Response constructor
, addPermissionResponse
) where
import Network.AWS.Prelude
import Network.AWS.Request.Query
import Network.AWS.SNS.Types
import qualified GHC.Exts
data AddPermission = AddPermission
{ _apAWSAccountId :: List "member" Text
, _apActionName :: List "member" Text
, _apLabel :: Text
, _apTopicArn :: Text
} deriving (Eq, Ord, Read, Show)
-- | 'AddPermission' constructor.
--
-- The fields accessible through corresponding lenses are:
--
-- * 'apAWSAccountId' @::@ ['Text']
--
-- * 'apActionName' @::@ ['Text']
--
-- * 'apLabel' @::@ 'Text'
--
-- * 'apTopicArn' @::@ 'Text'
--
addPermission :: Text -- ^ 'apTopicArn'
-> Text -- ^ 'apLabel'
-> AddPermission
addPermission p1 p2 = AddPermission
{ _apTopicArn = p1
, _apLabel = p2
, _apAWSAccountId = mempty
, _apActionName = mempty
}
-- | The AWS account IDs of the users (principals) who will be given access to the
-- specified actions. The users must have AWS accounts, but do not need to be
-- signed up for this service.
apAWSAccountId :: Lens' AddPermission [Text]
apAWSAccountId = lens _apAWSAccountId (\s a -> s { _apAWSAccountId = a }) . _List
-- | The action you want to allow for the specified principal(s).
--
-- Valid values: any Amazon SNS action name.
apActionName :: Lens' AddPermission [Text]
apActionName = lens _apActionName (\s a -> s { _apActionName = a }) . _List
-- | A unique identifier for the new policy statement.
apLabel :: Lens' AddPermission Text
apLabel = lens _apLabel (\s a -> s { _apLabel = a })
-- | The ARN of the topic whose access control policy you wish to modify.
apTopicArn :: Lens' AddPermission Text
apTopicArn = lens _apTopicArn (\s a -> s { _apTopicArn = a })
data AddPermissionResponse = AddPermissionResponse
deriving (Eq, Ord, Read, Show, Generic)
-- | 'AddPermissionResponse' constructor.
addPermissionResponse :: AddPermissionResponse
addPermissionResponse = AddPermissionResponse
instance ToPath AddPermission where
toPath = const "/"
instance ToQuery AddPermission where
toQuery AddPermission{..} = mconcat
[ "AWSAccountId" =? _apAWSAccountId
, "ActionName" =? _apActionName
, "Label" =? _apLabel
, "TopicArn" =? _apTopicArn
]
instance ToHeaders AddPermission
instance AWSRequest AddPermission where
type Sv AddPermission = SNS
type Rs AddPermission = AddPermissionResponse
request = post "AddPermission"
response = nullResponse AddPermissionResponse
| kim/amazonka | amazonka-sns/gen/Network/AWS/SNS/AddPermission.hs | mpl-2.0 | 4,085 | 0 | 10 | 953 | 548 | 333 | 215 | 64 | 1 |
-------------------------------------------------------------------------------
-- |
-- Module : System.Hardware.Haskino.ShallowDeepPlugin.RepPushPass
-- Copyright : (c) University of Kansas
-- License : BSD3
-- Stability : experimental
--
-- Rep Push Pass
-- This pass is used to transform shallow expressions into the
-- deep expression language. It uses rules like the following:
--
-- forall (b1 :: Bool) (b2 :: Bool).
-- rep_ (b1 || b2)
-- =
-- (rep_ b1) ||* (rep_ b2)
--
-- Each of the from and to operations (in the example above '||' and
-- '||*' are specified in the xlatList data table.
-------------------------------------------------------------------------------
{-# LANGUAGE GeneralizedNewtypeDeriving #-}
{-# LANGUAGE TemplateHaskell #-}
module System.Hardware.Haskino.ShallowDeepPlugin.RepPushPass (repPushPass) where
import Control.Monad.Reader
import CoreMonad
import Data.Bits as DB
import Data.Boolean
import Data.Boolean.Numbers as BN
import Data.Boolean.Bits as BB
import Data.List
import GhcPlugins
import System.Hardware.Haskino
import System.Hardware.Haskino.ShallowDeepPlugin.Utils
data XlatEntry = XlatEntry { fromId :: BindM Id
, toId :: BindM Id
}
-- The following talbe defines the names of the Shallow DSL functions
-- to translate from and the Deep DSL functions to translate to.
xlatList :: [XlatEntry]
xlatList = [ XlatEntry (thNameToId 'not)
(thNameToId 'Data.Boolean.notB)
, XlatEntry (thNameToId '(||))
(thNameToId '(||*))
, XlatEntry (thNameToId '(&&))
(thNameToId '(&&*))
, XlatEntry (thNameToId '(==))
(thNameToId 'eqE)
, XlatEntry (thNameToId '(/=))
(thNameToId 'neqE)
, XlatEntry (thNameToId '(>))
(thNameToId 'greatE)
, XlatEntry (thNameToId '(<))
(thNameToId 'lessE)
, XlatEntry (thNameToId '(>=))
(thNameToId 'greateqE)
, XlatEntry (thNameToId '(<=))
(thNameToId 'lesseqE)
, XlatEntry (thNameToId '(+))
(thNameToId '(+))
, XlatEntry (thNameToId '(-))
(thNameToId '(-))
, XlatEntry (thNameToId '(/))
(thNameToId '(/))
, XlatEntry (thNameToId '(*))
(thNameToId '(*))
, XlatEntry (thNameToId 'Prelude.div)
(thNameToId 'BN.div)
, XlatEntry (thNameToId 'Prelude.rem)
(thNameToId 'BN.rem)
, XlatEntry (thNameToId 'Prelude.quot)
(thNameToId 'BN.quot)
, XlatEntry (thNameToId 'Prelude.mod)
(thNameToId 'BN.mod)
, XlatEntry (thNameToId 'negate)
(thNameToId 'negate)
, XlatEntry (thNameToId 'abs)
(thNameToId 'abs)
, XlatEntry (thNameToId 'signum)
(thNameToId 'signum)
, XlatEntry (thNameToId '(DB..&.))
(thNameToId '(BB..&.))
, XlatEntry (thNameToId '(DB..|.))
(thNameToId '(BB..|.))
, XlatEntry (thNameToId 'DB.xor)
(thNameToId 'BB.xor)
, XlatEntry (thNameToId 'DB.complement)
(thNameToId 'BB.complement)
, XlatEntry (thNameToId 'DB.setBit)
(thNameToId 'BB.setBit)
, XlatEntry (thNameToId 'DB.clearBit)
(thNameToId 'BB.clearBit)
, XlatEntry (thNameToId 'DB.testBit)
(thNameToId 'BB.testBit)
, XlatEntry (thNameToId 'DB.finiteBitSize)
(thNameToId 'BB.bitSize)
, XlatEntry (thNameToId 'DB.isSigned)
(thNameToId 'BB.isSigned)
, XlatEntry (thNameToId 'DB.shiftL)
(thNameToId 'BB.shiftL)
, XlatEntry (thNameToId 'DB.shiftR)
(thNameToId 'BB.shiftR)
, XlatEntry (thNameToId 'DB.rotateL)
(thNameToId 'BB.rotateL)
, XlatEntry (thNameToId 'DB.rotateR)
(thNameToId 'BB.rotateR)
, XlatEntry (thNameToId '(++))
(thNameToId '(++*))
, XlatEntry (thNameToId '(:))
(thNameToId '(*:))
, XlatEntry (thNameToId '(!!))
(thNameToId '(!!*))
, XlatEntry (thNameToId 'head)
(thNameToId 'headE)
, XlatEntry (thNameToId 'tail)
(thNameToId 'tailE)
, XlatEntry (thNameToId 'length)
(thNameToId 'len)
, XlatEntry (thNameToId 'drop)
(thNameToId 'dropE)
, XlatEntry (thNameToId 'take)
(thNameToId 'takeE)
, XlatEntry (thNameToId 'Data.List.reverse)
(thNameToId 'reverseE)
, XlatEntry (thNameToId 'null)
(thNameToId 'nullE)
, XlatEntry (thNameToId 'showB)
(thNameToId 'showE)
, XlatEntry (thNameToId 'fromIntegral)
(thNameToId 'fromIntegralE)
]
-- TBD add floating to above
data BindEnv
= BindEnv
{ pluginModGuts :: ModGuts
}
newtype BindM a = BindM { runBindM :: ReaderT BindEnv CoreM a }
deriving (Functor, Applicative, Monad
,MonadIO, MonadReader BindEnv)
instance PassCoreM BindM where
liftCoreM = BindM . ReaderT . const
getModGuts = BindM $ ReaderT (return . pluginModGuts)
repPushPass :: ModGuts -> CoreM ModGuts
repPushPass guts = do
bindsOnlyPass (\x -> (runReaderT (runBindM $ (mapM changeRep) x) (BindEnv guts))) guts
changeRep :: CoreBind -> BindM CoreBind
changeRep (NonRec b e) = do
let (bs, e') = collectBinders e
e'' <- changeRepExpr e'
let e''' = mkLams bs e''
return (NonRec b e''')
changeRep (Rec bs) = do
bs' <- changeRep' bs
return $ Rec bs'
changeRep' :: [(Id, CoreExpr)] -> BindM [(Id, CoreExpr)]
changeRep' [] = return []
changeRep' ((b, e) : bs) = do
let (lbs, e') = collectBinders e
e'' <- changeRepExpr e'
let e''' = mkLams lbs e''
bs' <- changeRep' bs
return $ (b, e''') : bs'
changeRepExpr :: CoreExpr -> BindM CoreExpr
changeRepExpr e = do
repId <- thNameToId repNameTH
case e of
Var v -> return $ Var v
Lit l -> return $ Lit l
Type ty -> return $ Type ty
Coercion co -> return $ Coercion co
App e1 e2 -> do
let (b, args) = collectArgs e
let defaultReturn = do
e1' <- changeRepExpr e1
e2' <- changeRepExpr e2
return $ App e1' e2'
case b of
Var v | v == repId -> do
case args of
[_ty, _dict, e'] -> do
let (b', args') = collectArgs e'
case b' of
Var v' -> do
inList <- funcInXlatList v'
case inList of
Just xe -> pushRep xe args' (exprType e)
_ -> defaultReturn
_ -> defaultReturn
_ -> defaultReturn
_ -> defaultReturn
Lam tb el -> do
e' <- changeRepExpr el
return $ Lam tb e'
Let bind body -> do
body' <- changeRepExpr body
bind' <- case bind of
(NonRec v el) -> do
e' <- changeRepExpr el
return $ NonRec v e'
(Rec rbs) -> do
rbs' <- changeRepExpr' rbs
return $ Rec rbs'
return $ Let bind' body'
Case ec tb ty alts -> do
e' <- changeRepExpr ec
alts' <- changeRepExprAlts alts
return $ Case e' tb ty alts'
Tick t et -> do
e' <- changeRepExpr et
return $ Tick t e'
Cast ec co -> do
e' <- changeRepExpr ec
return $ Cast e' co
changeRepExpr' :: [(Id, CoreExpr)] -> BindM [(Id, CoreExpr)]
changeRepExpr' [] = return []
changeRepExpr' ((b, e) : bs) = do
e' <- changeRepExpr e
bs' <- changeRepExpr' bs
return $ (b, e') : bs'
changeRepExprAlts :: [GhcPlugins.Alt CoreBndr] -> BindM [GhcPlugins.Alt CoreBndr]
changeRepExprAlts [] = return []
changeRepExprAlts ((ac, b, a) : as) = do
a' <- changeRepExpr a
bs' <- changeRepExprAlts as
return $ (ac, b, a') : bs'
funcInXlatList :: Id -> BindM (Maybe XlatEntry)
funcInXlatList idf = do
funcInXlatList' idf xlatList
where
funcInXlatList' :: Id -> [XlatEntry] -> BindM (Maybe XlatEntry)
funcInXlatList' _ [] = return Nothing
funcInXlatList' idf' (xl:xls) = do
fId <- fromId xl
if fId == idf'
then return $ Just xl
else funcInXlatList' idf' xls
pushRep :: XlatEntry -> [CoreExpr] -> Type -> BindM CoreExpr
pushRep xe args origRTy = do
fi <- fromId xe
ti <- toId xe
-- Break down the arguments from the old function
-- into foralls, args, and return types.
let (_fromForAlls, fromFuncTy) = splitForAllTys $ idType fi
let (_fromArgTys, fromRetTy) = splitFunTys fromFuncTy
let (toForAlls, toFuncTy) = splitForAllTys $ idType ti
let (toArgTys, toRetTy) = splitFunTys toFuncTy
-- Get the count of non-dictionary args in the new function
let argCount = countNonDictTypes toArgTys
let dictCount = (length toArgTys) - argCount
-- Get the original args based on the argCount
let someArgs = take ((length args) - argCount) args
let origArgs = drop ((length args) - argCount) args
let dictTys = take dictCount toArgTys
let nonDictTys = drop dictCount toArgTys
let typeArgs = genForAllArgs toForAlls nonDictTys fromRetTy origArgs
exprTypeArgs <- mapM (thNameTyToTyConApp exprTyConTH) typeArgs
let exprTypeVars = map Type exprTypeArgs
dictArgs <- genDictArgs dictTys nonDictTys fromRetTy toRetTy origArgs origRTy
repArgs <- mapM repExpr origArgs
repArgs' <- mapM changeRepExpr repArgs
return $ mkCoreApps (Var ti) (exprTypeVars ++ dictArgs ++ repArgs')
genDictArgs :: [Type] -> [Type] -> Type -> Type -> [CoreExpr] -> Type -> BindM [CoreExpr]
genDictArgs [] _ _ _ _ _ = return []
genDictArgs (dty:dtys) tys frty trty args orty = do
let (tyConTy, ty') = splitAppTys dty
dictTys <- mapM findTypeMatch ty'
dict <- buildDictionaryTyConTs (tyConAppTyCon tyConTy) dictTys
dicts <- genDictArgs dtys tys frty trty args orty
return $ dict:dicts
where
findTypeMatch :: Type -> BindM Type
findTypeMatch fty =
case findIndex (typeIn fty) tys of
Just idx -> do
-- Find the index of the from function arg which matches the
-- type required by the dictionary
let fromArgTy = exprType $ args !! idx
-- Get the base type only of the from type, removing any
-- Expr if it already exists.
let dictTy = case splitTyConApp_maybe fromArgTy of
Just (_, [ty']) -> ty'
_ -> fromArgTy
-- Get the type of the to function arg at the same index
let toArgTy = tys !! idx
-- Determine if it has a type constructor
let tys_m = splitTyConApp_maybe toArgTy
case tys_m of
-- If it does, we are dealing with a function
-- of type Class a => Expr a -> ... -> retType
-- so the dictionary type is of a not Expr a
Just (_tyCon, _tys') -> return dictTy
-- If there is no TyCon, then we have a
-- function of type Class a => a -> ... -> retType
-- so the dictionary type needs to be Expr a.
Nothing -> thNameTyToTyConApp exprTyConTH dictTy
Nothing -> do
-- Determine if the return type has a type constructor
let tys_m = splitTyConApp_maybe trty
-- Get the base type only of the from type, removing any
-- Expr if it already exists.
let orty' = case splitTyConApp_maybe orty of
Just (_, [ty']) -> ty'
_ -> orty
case tys_m of
-- If it does, we are dealing with a function
-- of type Class a => Expr a -> ... -> Expr retType
-- so the dictionary type is of a not Expr a
Just (_tyCon, _tys') -> return orty'
-- If there is no TyCon, then we have a
-- function of type Class a => a -> ... -> retType
-- so the dictionary type needs to be Expr a.
Nothing -> thNameTyToTyConApp exprTyConTH orty'
genForAllArgs :: [TyVar] -> [Type] -> Type -> [CoreExpr] -> [Type]
genForAllArgs [] _ _ _ = []
genForAllArgs (tv:tvs) tys rty args =
case findIndex (eqType (mkTyVarTy tv)) tys of
Just idx -> (exprType $ args !! idx) : genForAllArgs tvs tys rty args
Nothing -> rty : genForAllArgs tvs tys rty args
countNonDictTypes :: [Type] -> Int
countNonDictTypes [] = 0
countNonDictTypes (ty:tys) = if isDictTy ty
then countNonDictTypes tys
else 1 + countNonDictTypes tys
typeIn :: Type -> Type -> Bool
typeIn t ty =
if t `eqType` ty
then True
else
let (_, tys') = splitTyConApp ty in
t `eqType` last tys'
| ku-fpg/kansas-amber | System/Hardware/Haskino/ShallowDeepPlugin/RepPushPass.hs | bsd-3-clause | 13,617 | 0 | 33 | 4,767 | 3,701 | 1,918 | 1,783 | 275 | 15 |
{-# LANGUAGE CPP #-}
module RnSplice (
rnTopSpliceDecls,
rnSpliceType, rnSpliceExpr, rnSplicePat, rnSpliceDecl,
rnBracket,
checkThLocalName
#ifdef GHCI
, traceSplice, SpliceInfo(..)
#endif
) where
#include "HsVersions.h"
import Name
import NameSet
import HsSyn
import RdrName
import TcRnMonad
import Kind
import RnEnv
import RnSource ( rnSrcDecls, findSplice )
import RnPat ( rnPat )
import BasicTypes ( TopLevelFlag, isTopLevel )
import Outputable
import Module
import SrcLoc
import RnTypes ( rnLHsType )
import Control.Monad ( unless, when )
import {-# SOURCE #-} RnExpr ( rnLExpr )
import TcEnv ( checkWellStaged )
import THNames ( liftName )
#ifdef GHCI
import DynFlags
import FastString
import ErrUtils ( dumpIfSet_dyn_printer )
import TcEnv ( tcMetaTy )
import Hooks
import Var ( Id )
import THNames ( quoteExpName, quotePatName, quoteDecName, quoteTypeName
, decsQTyConName, expQTyConName, patQTyConName, typeQTyConName, )
import {-# SOURCE #-} TcExpr ( tcPolyExpr )
import {-# SOURCE #-} TcSplice ( runMetaD, runMetaE, runMetaP, runMetaT, tcTopSpliceExpr )
#endif
import qualified GHC.LanguageExtensions as LangExt
{-
************************************************************************
* *
Template Haskell brackets
* *
************************************************************************
-}
rnBracket :: HsExpr RdrName -> HsBracket RdrName -> RnM (HsExpr Name, FreeVars)
rnBracket e br_body
= addErrCtxt (quotationCtxtDoc br_body) $
do { -- Check that -XTemplateHaskellQuotes is enabled and available
thQuotesEnabled <- xoptM LangExt.TemplateHaskellQuotes
; unless thQuotesEnabled $
failWith ( vcat
[ text "Syntax error on" <+> ppr e
, text ("Perhaps you intended to use TemplateHaskell"
++ " or TemplateHaskellQuotes") ] )
-- Check for nested brackets
; cur_stage <- getStage
; case cur_stage of
{ Splice Typed -> checkTc (isTypedBracket br_body)
illegalUntypedBracket
; Splice Untyped -> checkTc (not (isTypedBracket br_body))
illegalTypedBracket
; Comp -> return ()
; Brack {} -> failWithTc illegalBracket
}
-- Brackets are desugared to code that mentions the TH package
; recordThUse
; case isTypedBracket br_body of
True -> do { traceRn (text "Renaming typed TH bracket")
; (body', fvs_e) <-
setStage (Brack cur_stage RnPendingTyped) $
rn_bracket cur_stage br_body
; return (HsBracket body', fvs_e) }
False -> do { traceRn (text "Renaming untyped TH bracket")
; ps_var <- newMutVar []
; (body', fvs_e) <-
setStage (Brack cur_stage (RnPendingUntyped ps_var)) $
rn_bracket cur_stage br_body
; pendings <- readMutVar ps_var
; return (HsRnBracketOut body' pendings, fvs_e) }
}
rn_bracket :: ThStage -> HsBracket RdrName -> RnM (HsBracket Name, FreeVars)
rn_bracket outer_stage br@(VarBr flg rdr_name)
= do { name <- lookupOccRn rdr_name
; this_mod <- getModule
; when (flg && nameIsLocalOrFrom this_mod name) $
-- Type variables can be quoted in TH. See #5721.
do { mb_bind_lvl <- lookupLocalOccThLvl_maybe name
; case mb_bind_lvl of
{ Nothing -> return () -- Can happen for data constructors,
-- but nothing needs to be done for them
; Just (top_lvl, bind_lvl) -- See Note [Quoting names]
| isTopLevel top_lvl
-> when (isExternalName name) (keepAlive name)
| otherwise
-> do { traceRn (text "rn_bracket VarBr" <+> ppr name <+> ppr bind_lvl <+> ppr outer_stage)
; checkTc (thLevel outer_stage + 1 == bind_lvl)
(quotedNameStageErr br) }
}
}
; return (VarBr flg name, unitFV name) }
rn_bracket _ (ExpBr e) = do { (e', fvs) <- rnLExpr e
; return (ExpBr e', fvs) }
rn_bracket _ (PatBr p) = rnPat ThPatQuote p $ \ p' -> return (PatBr p', emptyFVs)
rn_bracket _ (TypBr t) = do { (t', fvs) <- rnLHsType TypBrCtx t
; return (TypBr t', fvs) }
rn_bracket _ (DecBrL decls)
= do { group <- groupDecls decls
; gbl_env <- getGblEnv
; let new_gbl_env = gbl_env { tcg_dus = emptyDUs }
-- The emptyDUs is so that we just collect uses for this
-- group alone in the call to rnSrcDecls below
; (tcg_env, group') <- setGblEnv new_gbl_env $
rnSrcDecls group
-- Discard the tcg_env; it contains only extra info about fixity
; traceRn (text "rn_bracket dec" <+> (ppr (tcg_dus tcg_env) $$
ppr (duUses (tcg_dus tcg_env))))
; return (DecBrG group', duUses (tcg_dus tcg_env)) }
where
groupDecls :: [LHsDecl RdrName] -> RnM (HsGroup RdrName)
groupDecls decls
= do { (group, mb_splice) <- findSplice decls
; case mb_splice of
{ Nothing -> return group
; Just (splice, rest) ->
do { group' <- groupDecls rest
; let group'' = appendGroups group group'
; return group'' { hs_splcds = noLoc splice : hs_splcds group' }
}
}}
rn_bracket _ (DecBrG _) = panic "rn_bracket: unexpected DecBrG"
rn_bracket _ (TExpBr e) = do { (e', fvs) <- rnLExpr e
; return (TExpBr e', fvs) }
quotationCtxtDoc :: HsBracket RdrName -> SDoc
quotationCtxtDoc br_body
= hang (text "In the Template Haskell quotation")
2 (ppr br_body)
illegalBracket :: SDoc
illegalBracket =
text "Template Haskell brackets cannot be nested" <+>
text "(without intervening splices)"
illegalTypedBracket :: SDoc
illegalTypedBracket =
text "Typed brackets may only appear in typed splices."
illegalUntypedBracket :: SDoc
illegalUntypedBracket =
text "Untyped brackets may only appear in untyped splices."
quotedNameStageErr :: HsBracket RdrName -> SDoc
quotedNameStageErr br
= sep [ text "Stage error: the non-top-level quoted name" <+> ppr br
, text "must be used at the same stage at which is is bound" ]
#ifndef GHCI
rnTopSpliceDecls :: HsSplice RdrName -> RnM ([LHsDecl RdrName], FreeVars)
rnTopSpliceDecls e = failTH e "Template Haskell top splice"
rnSpliceType :: HsSplice RdrName -> PostTc Name Kind
-> RnM (HsType Name, FreeVars)
rnSpliceType e _ = failTH e "Template Haskell type splice"
rnSpliceExpr :: HsSplice RdrName -> RnM (HsExpr Name, FreeVars)
rnSpliceExpr e = failTH e "Template Haskell splice"
rnSplicePat :: HsSplice RdrName -> RnM (Either (Pat RdrName) (Pat Name), FreeVars)
rnSplicePat e = failTH e "Template Haskell pattern splice"
rnSpliceDecl :: SpliceDecl RdrName -> RnM (SpliceDecl Name, FreeVars)
rnSpliceDecl e = failTH e "Template Haskell declaration splice"
#else
{-
*********************************************************
* *
Splices
* *
*********************************************************
Note [Free variables of typed splices]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Consider renaming this:
f = ...
h = ...$(thing "f")...
where the splice is a *typed* splice. The splice can expand into
literally anything, so when we do dependency analysis we must assume
that it might mention 'f'. So we simply treat all locally-defined
names as mentioned by any splice. This is terribly brutal, but I
don't see what else to do. For example, it'll mean that every
locally-defined thing will appear to be used, so no unused-binding
warnings. But if we miss the dependency, then we might typecheck 'h'
before 'f', and that will crash the type checker because 'f' isn't in
scope.
Currently, I'm not treating a splice as also mentioning every import,
which is a bit inconsistent -- but there are a lot of them. We might
thereby get some bogus unused-import warnings, but we won't crash the
type checker. Not very satisfactory really.
Note [Renamer errors]
~~~~~~~~~~~~~~~~~~~~~
It's important to wrap renamer calls in checkNoErrs, because the
renamer does not fail for out of scope variables etc. Instead it
returns a bogus term/type, so that it can report more than one error.
We don't want the type checker to see these bogus unbound variables.
-}
rnSpliceGen :: (HsSplice Name -> RnM (a, FreeVars)) -- Outside brackets, run splice
-> (HsSplice Name -> (PendingRnSplice, a)) -- Inside brackets, make it pending
-> HsSplice RdrName
-> RnM (a, FreeVars)
rnSpliceGen run_splice pend_splice splice
= addErrCtxt (spliceCtxt splice) $ do
{ stage <- getStage
; case stage of
Brack pop_stage RnPendingTyped
-> do { checkTc is_typed_splice illegalUntypedSplice
; (splice', fvs) <- setStage pop_stage $
rnSplice splice
; let (_pending_splice, result) = pend_splice splice'
; return (result, fvs) }
Brack pop_stage (RnPendingUntyped ps_var)
-> do { checkTc (not is_typed_splice) illegalTypedSplice
; (splice', fvs) <- setStage pop_stage $
rnSplice splice
; let (pending_splice, result) = pend_splice splice'
; ps <- readMutVar ps_var
; writeMutVar ps_var (pending_splice : ps)
; return (result, fvs) }
_ -> do { (splice', fvs1) <- checkNoErrs $
setStage (Splice splice_type) $
rnSplice splice
-- checkNoErrs: don't attempt to run the splice if
-- renaming it failed; otherwise we get a cascade of
-- errors from e.g. unbound variables
; (result, fvs2) <- run_splice splice'
; return (result, fvs1 `plusFV` fvs2) } }
where
is_typed_splice = isTypedSplice splice
splice_type = if is_typed_splice
then Typed
else Untyped
------------------
runRnSplice :: UntypedSpliceFlavour
-> (LHsExpr Id -> TcRn res)
-> (res -> SDoc) -- How to pretty-print res
-- Usually just ppr, but not for [Decl]
-> HsSplice Name -- Always untyped
-> TcRn res
runRnSplice flavour run_meta ppr_res splice
= do { splice' <- getHooked runRnSpliceHook return >>= ($ splice)
; let the_expr = case splice' of
HsUntypedSplice _ e -> e
HsQuasiQuote _ q qs str -> mkQuasiQuoteExpr flavour q qs str
HsTypedSplice {} -> pprPanic "runRnSplice" (ppr splice)
-- Typecheck the expression
; meta_exp_ty <- tcMetaTy meta_ty_name
; zonked_q_expr <- tcTopSpliceExpr Untyped $
tcPolyExpr the_expr meta_exp_ty
-- Run the expression
; result <- run_meta zonked_q_expr
; traceSplice (SpliceInfo { spliceDescription = what
, spliceIsDecl = is_decl
, spliceSource = Just the_expr
, spliceGenerated = ppr_res result })
; return result }
where
meta_ty_name = case flavour of
UntypedExpSplice -> expQTyConName
UntypedPatSplice -> patQTyConName
UntypedTypeSplice -> typeQTyConName
UntypedDeclSplice -> decsQTyConName
what = case flavour of
UntypedExpSplice -> "expression"
UntypedPatSplice -> "pattern"
UntypedTypeSplice -> "type"
UntypedDeclSplice -> "declarations"
is_decl = case flavour of
UntypedDeclSplice -> True
_ -> False
------------------
makePending :: UntypedSpliceFlavour
-> HsSplice Name
-> PendingRnSplice
makePending flavour (HsUntypedSplice n e)
= PendingRnSplice flavour n e
makePending flavour (HsQuasiQuote n quoter q_span quote)
= PendingRnSplice flavour n (mkQuasiQuoteExpr flavour quoter q_span quote)
makePending _ splice@(HsTypedSplice {})
= pprPanic "makePending" (ppr splice)
------------------
mkQuasiQuoteExpr :: UntypedSpliceFlavour -> Name -> SrcSpan -> FastString -> LHsExpr Name
-- Return the expression (quoter "...quote...")
-- which is what we must run in a quasi-quote
mkQuasiQuoteExpr flavour quoter q_span quote
= L q_span $ HsApp (L q_span $
HsApp (L q_span (HsVar (L q_span quote_selector)))
quoterExpr)
quoteExpr
where
quoterExpr = L q_span $! HsVar $! (L q_span quoter)
quoteExpr = L q_span $! HsLit $! HsString "" quote
quote_selector = case flavour of
UntypedExpSplice -> quoteExpName
UntypedPatSplice -> quotePatName
UntypedTypeSplice -> quoteTypeName
UntypedDeclSplice -> quoteDecName
---------------------
rnSplice :: HsSplice RdrName -> RnM (HsSplice Name, FreeVars)
-- Not exported...used for all
rnSplice (HsTypedSplice splice_name expr)
= do { checkTH expr "Template Haskell typed splice"
; loc <- getSrcSpanM
; n' <- newLocalBndrRn (L loc splice_name)
; (expr', fvs) <- rnLExpr expr
; return (HsTypedSplice n' expr', fvs) }
rnSplice (HsUntypedSplice splice_name expr)
= do { checkTH expr "Template Haskell untyped splice"
; loc <- getSrcSpanM
; n' <- newLocalBndrRn (L loc splice_name)
; (expr', fvs) <- rnLExpr expr
; return (HsUntypedSplice n' expr', fvs) }
rnSplice (HsQuasiQuote splice_name quoter q_loc quote)
= do { checkTH quoter "Template Haskell quasi-quote"
; loc <- getSrcSpanM
; splice_name' <- newLocalBndrRn (L loc splice_name)
-- Rename the quoter; akin to the HsVar case of rnExpr
; quoter' <- lookupOccRn quoter
; this_mod <- getModule
; when (nameIsLocalOrFrom this_mod quoter') $
checkThLocalName quoter'
; return (HsQuasiQuote splice_name' quoter' q_loc quote, unitFV quoter') }
---------------------
rnSpliceExpr :: HsSplice RdrName -> RnM (HsExpr Name, FreeVars)
rnSpliceExpr splice
= rnSpliceGen run_expr_splice pend_expr_splice splice
where
pend_expr_splice :: HsSplice Name -> (PendingRnSplice, HsExpr Name)
pend_expr_splice rn_splice
= (makePending UntypedExpSplice rn_splice, HsSpliceE rn_splice)
run_expr_splice :: HsSplice Name -> RnM (HsExpr Name, FreeVars)
run_expr_splice rn_splice
| isTypedSplice rn_splice -- Run it later, in the type checker
= do { -- Ugh! See Note [Splices] above
traceRn (text "rnSpliceExpr: typed expression splice")
; lcl_rdr <- getLocalRdrEnv
; gbl_rdr <- getGlobalRdrEnv
; let gbl_names = mkNameSet [gre_name gre | gre <- globalRdrEnvElts gbl_rdr
, isLocalGRE gre]
lcl_names = mkNameSet (localRdrEnvElts lcl_rdr)
; return (HsSpliceE rn_splice, lcl_names `plusFV` gbl_names) }
| otherwise -- Run it here, see Note [Running splices in the Renamer]
= do { traceRn (text "rnSpliceExpr: untyped expression splice")
; rn_expr <- runRnSplice UntypedExpSplice runMetaE ppr rn_splice
; (lexpr3, fvs) <- checkNoErrs (rnLExpr rn_expr)
; return (HsPar lexpr3, fvs) }
{- Note [Running splices in the Renamer]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Splices used to be run in the typechecker, which led to (Trac #4364). Since the
renamer must decide which expressions depend on which others, and it cannot
reliably do this for arbitrary splices, we used to conservatively say that
splices depend on all other expressions in scope. Unfortunately, this led to
the problem of cyclic type declarations seen in (Trac #4364). Instead, by
running splices in the renamer, we side-step the problem of determining
dependencies: by the time the dependency analysis happens, any splices have
already been run, and expression dependencies can be determined as usual.
However, see (Trac #9813), for an example where we would like to run splices
*after* performing dependency analysis (that is, after renaming). It would be
desirable to typecheck "non-splicy" expressions (those expressions that do not
contain splices directly or via dependence on an expression that does) before
"splicy" expressions, such that types/expressions within the same declaration
group would be available to `reify` calls, for example consider the following:
> module M where
> data D = C
> f = 1
> g = $(mapM reify ['f, 'D, ''C] ...)
Compilation of this example fails since D/C/f are not in the type environment
and thus cannot be reified as they have not been typechecked by the time the
splice is renamed and thus run.
These requirements are at odds: we do not want to run splices in the renamer as
we wish to first determine dependencies and typecheck certain expressions,
making them available to reify, but cannot accurately determine dependencies
without running splices in the renamer!
Indeed, the conclusion of (Trac #9813) was that it is not worth the complexity
to try and
a) implement and maintain the code for renaming/typechecking non-splicy
expressions before splicy expressions,
b) explain to TH users which expressions are/not available to reify at any
given point.
-}
----------------------
rnSpliceType :: HsSplice RdrName -> PostTc Name Kind
-> RnM (HsType Name, FreeVars)
rnSpliceType splice k
= rnSpliceGen run_type_splice pend_type_splice splice
where
pend_type_splice rn_splice
= (makePending UntypedTypeSplice rn_splice, HsSpliceTy rn_splice k)
run_type_splice rn_splice
= do { traceRn (text "rnSpliceType: untyped type splice")
; hs_ty2 <- runRnSplice UntypedTypeSplice runMetaT ppr rn_splice
; (hs_ty3, fvs) <- do { let doc = SpliceTypeCtx hs_ty2
; checkNoErrs $ rnLHsType doc hs_ty2 }
-- checkNoErrs: see Note [Renamer errors]
; return (HsParTy hs_ty3, fvs) }
-- Wrap the result of the splice in parens so that we don't
-- lose the outermost location set by runQuasiQuote (#7918)
{- Note [Partial Type Splices]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Partial Type Signatures are partially supported in TH type splices: only
anonymous wild cards are allowed.
-- ToDo: SLPJ says: I don't understand all this
Normally, named wild cards are collected before renaming a (partial) type
signature. However, TH type splices are run during renaming, i.e. after the
initial traversal, leading to out of scope errors for named wild cards. We
can't just extend the initial traversal to collect the named wild cards in TH
type splices, as we'd need to expand them, which is supposed to happen only
once, during renaming.
Similarly, the extra-constraints wild card is handled right before renaming
too, and is therefore also not supported in a TH type splice. Another reason
to forbid extra-constraints wild cards in TH type splices is that a single
signature can contain many TH type splices, whereas it mustn't contain more
than one extra-constraints wild card. Enforcing would this be hard the way
things are currently organised.
Anonymous wild cards pose no problem, because they start out without names and
are given names during renaming. These names are collected right after
renaming. The names generated for anonymous wild cards in TH type splices will
thus be collected as well.
For more details about renaming wild cards, see RnTypes.rnHsSigWcType
Note that partial type signatures are fully supported in TH declaration
splices, e.g.:
[d| foo :: _ => _
foo x y = x == y |]
This is because in this case, the partial type signature can be treated as a
whole signature, instead of as an arbitrary type.
-}
----------------------
-- | Rename a splice pattern. See Note [rnSplicePat]
rnSplicePat :: HsSplice RdrName -> RnM ( Either (Pat RdrName) (Pat Name)
, FreeVars)
rnSplicePat splice
= rnSpliceGen run_pat_splice pend_pat_splice splice
where
pend_pat_splice rn_splice
= (makePending UntypedPatSplice rn_splice, Right (SplicePat rn_splice))
run_pat_splice rn_splice
= do { traceRn (text "rnSplicePat: untyped pattern splice")
; pat <- runRnSplice UntypedPatSplice runMetaP ppr rn_splice
; return (Left (ParPat pat), emptyFVs) }
-- Wrap the result of the quasi-quoter in parens so that we don't
-- lose the outermost location set by runQuasiQuote (#7918)
----------------------
rnSpliceDecl :: SpliceDecl RdrName -> RnM (SpliceDecl Name, FreeVars)
rnSpliceDecl (SpliceDecl (L loc splice) flg)
= rnSpliceGen run_decl_splice pend_decl_splice splice
where
pend_decl_splice rn_splice
= (makePending UntypedDeclSplice rn_splice, SpliceDecl (L loc rn_splice) flg)
run_decl_splice rn_splice = pprPanic "rnSpliceDecl" (ppr rn_splice)
rnTopSpliceDecls :: HsSplice RdrName -> RnM ([LHsDecl RdrName], FreeVars)
-- Declaration splice at the very top level of the module
rnTopSpliceDecls splice
= do { (rn_splice, fvs) <- setStage (Splice Untyped) $
rnSplice splice
; traceRn (text "rnTopSpliceDecls: untyped declaration splice")
; decls <- runRnSplice UntypedDeclSplice runMetaD ppr_decls rn_splice
; return (decls,fvs) }
where
ppr_decls :: [LHsDecl RdrName] -> SDoc
ppr_decls ds = vcat (map ppr ds)
{-
Note [rnSplicePat]
~~~~~~~~~~~~~~~~~~
Renaming a pattern splice is a bit tricky, because we need the variables
bound in the pattern to be in scope in the RHS of the pattern. This scope
management is effectively done by using continuation-passing style in
RnPat, through the CpsRn monad. We don't wish to be in that monad here
(it would create import cycles and generally conflict with renaming other
splices), so we really want to return a (Pat RdrName) -- the result of
running the splice -- which can then be further renamed in RnPat, in
the CpsRn monad.
The problem is that if we're renaming a splice within a bracket, we
*don't* want to run the splice now. We really do just want to rename
it to an HsSplice Name. Of course, then we can't know what variables
are bound within the splice. So we accept any unbound variables and
rename them again when the bracket is spliced in. If a variable is brought
into scope by a pattern splice all is fine. If it is not then an error is
reported.
In any case, when we're done in rnSplicePat, we'll either have a
Pat RdrName (the result of running a top-level splice) or a Pat Name
(the renamed nested splice). Thus, the awkward return type of
rnSplicePat.
-}
spliceCtxt :: HsSplice RdrName -> SDoc
spliceCtxt splice
= hang (text "In the" <+> what) 2 (ppr splice)
where
what = case splice of
HsUntypedSplice {} -> text "untyped splice:"
HsTypedSplice {} -> text "typed splice:"
HsQuasiQuote {} -> text "quasi-quotation:"
-- | The splice data to be logged
data SpliceInfo
= SpliceInfo
{ spliceDescription :: String
, spliceSource :: Maybe (LHsExpr Name) -- Nothing <=> top-level decls
-- added by addTopDecls
, spliceIsDecl :: Bool -- True <=> put the generate code in a file
-- when -dth-dec-file is on
, spliceGenerated :: SDoc
}
-- Note that 'spliceSource' is *renamed* but not *typechecked*
-- Reason (a) less typechecking crap
-- (b) data constructors after type checking have been
-- changed to their *wrappers*, and that makes them
-- print always fully qualified
-- | outputs splice information for 2 flags which have different output formats:
-- `-ddump-splices` and `-dth-dec-file`
traceSplice :: SpliceInfo -> TcM ()
traceSplice (SpliceInfo { spliceDescription = sd, spliceSource = mb_src
, spliceGenerated = gen, spliceIsDecl = is_decl })
= do { loc <- case mb_src of
Nothing -> getSrcSpanM
Just (L loc _) -> return loc
; traceOptTcRn Opt_D_dump_splices (spliceDebugDoc loc)
; when is_decl $ -- Raw material for -dth-dec-file
do { dflags <- getDynFlags
; liftIO $ dumpIfSet_dyn_printer alwaysQualify dflags Opt_D_th_dec_file
(spliceCodeDoc loc) } }
where
-- `-ddump-splices`
spliceDebugDoc :: SrcSpan -> SDoc
spliceDebugDoc loc
= let code = case mb_src of
Nothing -> ending
Just e -> nest 2 (ppr e) : ending
ending = [ text "======>", nest 2 gen ]
in hang (ppr loc <> colon <+> text "Splicing" <+> text sd)
2 (sep code)
-- `-dth-dec-file`
spliceCodeDoc :: SrcSpan -> SDoc
spliceCodeDoc loc
= vcat [ text "--" <+> ppr loc <> colon <+> text "Splicing" <+> text sd
, gen ]
illegalTypedSplice :: SDoc
illegalTypedSplice = text "Typed splices may not appear in untyped brackets"
illegalUntypedSplice :: SDoc
illegalUntypedSplice = text "Untyped splices may not appear in typed brackets"
-- spliceResultDoc :: OutputableBndr id => LHsExpr id -> SDoc
-- spliceResultDoc expr
-- = vcat [ hang (text "In the splice:")
-- 2 (char '$' <> pprParendExpr expr)
-- , text "To see what the splice expanded to, use -ddump-splices" ]
#endif
checkThLocalName :: Name -> RnM ()
checkThLocalName name
| isUnboundName name -- Do not report two errors for
= return () -- $(not_in_scope args)
| otherwise
= do { traceRn (text "checkThLocalName" <+> ppr name)
; mb_local_use <- getStageAndBindLevel name
; case mb_local_use of {
Nothing -> return () ; -- Not a locally-bound thing
Just (top_lvl, bind_lvl, use_stage) ->
do { let use_lvl = thLevel use_stage
; checkWellStaged (quotes (ppr name)) bind_lvl use_lvl
; traceRn (text "checkThLocalName" <+> ppr name <+> ppr bind_lvl <+> ppr use_stage <+> ppr use_lvl)
; checkCrossStageLifting top_lvl bind_lvl use_stage use_lvl name } } }
--------------------------------------
checkCrossStageLifting :: TopLevelFlag -> ThLevel -> ThStage -> ThLevel
-> Name -> TcM ()
-- We are inside brackets, and (use_lvl > bind_lvl)
-- Now we must check whether there's a cross-stage lift to do
-- Examples \x -> [| x |]
-- [| map |]
--
-- This code is similar to checkCrossStageLifting in TcExpr, but
-- this is only run on *untyped* brackets.
checkCrossStageLifting top_lvl bind_lvl use_stage use_lvl name
| Brack _ (RnPendingUntyped ps_var) <- use_stage -- Only for untyped brackets
, use_lvl > bind_lvl -- Cross-stage condition
= check_cross_stage_lifting top_lvl name ps_var
| otherwise
= return ()
check_cross_stage_lifting :: TopLevelFlag -> Name -> TcRef [PendingRnSplice] -> TcM ()
check_cross_stage_lifting top_lvl name ps_var
| isTopLevel top_lvl
-- Top-level identifiers in this module,
-- (which have External Names)
-- are just like the imported case:
-- no need for the 'lifting' treatment
-- E.g. this is fine:
-- f x = x
-- g y = [| f 3 |]
= when (isExternalName name) (keepAlive name)
-- See Note [Keeping things alive for Template Haskell]
| otherwise
= -- Nested identifiers, such as 'x' in
-- E.g. \x -> [| h x |]
-- We must behave as if the reference to x was
-- h $(lift x)
-- We use 'x' itself as the SplicePointName, used by
-- the desugarer to stitch it all back together.
-- If 'x' occurs many times we may get many identical
-- bindings of the same SplicePointName, but that doesn't
-- matter, although it's a mite untidy.
do { traceRn (text "checkCrossStageLifting" <+> ppr name)
-- Construct the (lift x) expression
; let lift_expr = nlHsApp (nlHsVar liftName) (nlHsVar name)
pend_splice = PendingRnSplice UntypedExpSplice name lift_expr
-- Update the pending splices
; ps <- readMutVar ps_var
; writeMutVar ps_var (pend_splice : ps) }
{-
Note [Keeping things alive for Template Haskell]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Consider
f x = x+1
g y = [| f 3 |]
Here 'f' is referred to from inside the bracket, which turns into data
and mentions only f's *name*, not 'f' itself. So we need some other
way to keep 'f' alive, lest it get dropped as dead code. That's what
keepAlive does. It puts it in the keep-alive set, which subsequently
ensures that 'f' stays as a top level binding.
This must be done by the renamer, not the type checker (as of old),
because the type checker doesn't typecheck the body of untyped
brackets (Trac #8540).
A thing can have a bind_lvl of outerLevel, but have an internal name:
foo = [d| op = 3
bop = op + 1 |]
Here the bind_lvl of 'op' is (bogusly) outerLevel, even though it is
bound inside a bracket. That is because we don't even even record
binding levels for top-level things; the binding levels are in the
LocalRdrEnv.
So the occurrence of 'op' in the rhs of 'bop' looks a bit like a
cross-stage thing, but it isn't really. And in fact we never need
to do anything here for top-level bound things, so all is fine, if
a bit hacky.
For these chaps (which have Internal Names) we don't want to put
them in the keep-alive set.
Note [Quoting names]
~~~~~~~~~~~~~~~~~~~~
A quoted name 'n is a bit like a quoted expression [| n |], except that we
have no cross-stage lifting (c.f. TcExpr.thBrackId). So, after incrementing
the use-level to account for the brackets, the cases are:
bind > use Error
bind = use+1 OK
bind < use
Imported things OK
Top-level things OK
Non-top-level Error
where 'use' is the binding level of the 'n quote. (So inside the implied
bracket the level would be use+1.)
Examples:
f 'map -- OK; also for top-level defns of this module
\x. f 'x -- Not ok (bind = 1, use = 1)
-- (whereas \x. f [| x |] might have been ok, by
-- cross-stage lifting
\y. [| \x. $(f 'y) |] -- Not ok (bind =1, use = 1)
[| \x. $(f 'x) |] -- OK (bind = 2, use = 1)
-}
| tjakway/ghcjvm | compiler/rename/RnSplice.hs | bsd-3-clause | 32,074 | 0 | 21 | 9,499 | 2,261 | 1,176 | 1,085 | 160 | 5 |
module Main (main) where
import qualified Distribution.ModuleName as ModuleName
import Distribution.PackageDescription
import Distribution.PackageDescription.Check hiding (doesFileExist)
import Distribution.PackageDescription.Configuration
import Distribution.PackageDescription.Parse
import Distribution.System
import Distribution.Simple
import Distribution.Simple.Configure
import Distribution.Simple.LocalBuildInfo
import Distribution.Simple.Program
import Distribution.Simple.Program.HcPkg
import Distribution.Simple.Utils (defaultPackageDesc, writeFileAtomic, toUTF8)
import Distribution.Simple.Build (writeAutogenFiles)
import Distribution.Simple.Register
import Distribution.Text
import Distribution.Verbosity
import qualified Distribution.InstalledPackageInfo as Installed
import qualified Distribution.Simple.PackageIndex as PackageIndex
import Control.Monad
import qualified Data.ByteString.Lazy.Char8 as BS
import Data.List
import Data.Maybe
import System.IO
import System.Directory
import System.Environment
import System.Exit
import System.FilePath
main :: IO ()
main = do hSetBuffering stdout LineBuffering
args <- getArgs
case args of
"hscolour" : dir : distDir : args' ->
runHsColour dir distDir args'
"check" : dir : [] ->
doCheck dir
"copy" : dir : distDir
: strip : myDestDir : myPrefix : myLibdir : myDocdir
: ghcLibWays : args' ->
doCopy dir distDir
strip myDestDir myPrefix myLibdir myDocdir
("dyn" `elem` words ghcLibWays)
args'
"register" : dir : distDir : ghc : ghcpkg : topdir
: myDestDir : myPrefix : myLibdir : myDocdir
: relocatableBuild : args' ->
doRegister dir distDir ghc ghcpkg topdir
myDestDir myPrefix myLibdir myDocdir
relocatableBuild args'
"configure" : dir : distDir : dll0Modules : config_args ->
generate dir distDir dll0Modules config_args
"sdist" : dir : distDir : [] ->
doSdist dir distDir
["--version"] ->
defaultMainArgs ["--version"]
_ -> die syntax_error
syntax_error :: [String]
syntax_error =
["syntax: ghc-cabal configure <configure-args> -- <distdir> <directory>...",
" ghc-cabal install <ghc-pkg> <directory> <distdir> <destdir> <prefix> <args>...",
" ghc-cabal hscolour <distdir> <directory> <args>..."]
die :: [String] -> IO a
die errs = do mapM_ (hPutStrLn stderr) errs
exitWith (ExitFailure 1)
-- XXX Should use bracket
withCurrentDirectory :: FilePath -> IO a -> IO a
withCurrentDirectory directory io
= do curDirectory <- getCurrentDirectory
setCurrentDirectory directory
r <- io
setCurrentDirectory curDirectory
return r
-- We need to use the autoconfUserHooks, as the packages that use
-- configure can create a .buildinfo file, and we need any info that
-- ends up in it.
userHooks :: UserHooks
userHooks = autoconfUserHooks
runDefaultMain :: IO ()
runDefaultMain
= do let verbosity = normal
gpdFile <- defaultPackageDesc verbosity
gpd <- readPackageDescription verbosity gpdFile
case buildType (flattenPackageDescription gpd) of
Just Configure -> defaultMainWithHooks autoconfUserHooks
-- time has a "Custom" Setup.hs, but it's actually Configure
-- plus a "./Setup test" hook. However, Cabal is also
-- "Custom", but doesn't have a configure script.
Just Custom ->
do configureExists <- doesFileExist "configure"
if configureExists
then defaultMainWithHooks autoconfUserHooks
else defaultMain
-- not quite right, but good enough for us:
_ -> defaultMain
doSdist :: FilePath -> FilePath -> IO ()
doSdist directory distDir
= withCurrentDirectory directory
$ withArgs (["sdist", "--builddir", distDir])
runDefaultMain
doCheck :: FilePath -> IO ()
doCheck directory
= withCurrentDirectory directory
$ do let verbosity = normal
gpdFile <- defaultPackageDesc verbosity
gpd <- readPackageDescription verbosity gpdFile
case partition isFailure $ checkPackage gpd Nothing of
([], []) -> return ()
([], warnings) -> mapM_ print warnings
(errs, _) -> do mapM_ print errs
exitWith (ExitFailure 1)
where isFailure (PackageDistSuspicious {}) = False
isFailure _ = True
runHsColour :: FilePath -> FilePath -> [String] -> IO ()
runHsColour directory distdir args
= withCurrentDirectory directory
$ defaultMainArgs ("hscolour" : "--builddir" : distdir : args)
doCopy :: FilePath -> FilePath
-> FilePath -> FilePath -> FilePath -> FilePath -> FilePath -> Bool
-> [String]
-> IO ()
doCopy directory distDir
strip myDestDir myPrefix myLibdir myDocdir withSharedLibs
args
= withCurrentDirectory directory $ do
let copyArgs = ["copy", "--builddir", distDir]
++ (if null myDestDir
then []
else ["--destdir", myDestDir])
++ args
copyHooks = userHooks {
copyHook = noGhcPrimHook
$ modHook False
$ copyHook userHooks
}
defaultMainWithHooksArgs copyHooks copyArgs
where
noGhcPrimHook f pd lbi us flags
= let pd'
| packageName pd == PackageName "ghc-prim" =
case library pd of
Just lib ->
let ghcPrim = fromJust (simpleParse "GHC.Prim")
ems = filter (ghcPrim /=) (exposedModules lib)
lib' = lib { exposedModules = ems }
in pd { library = Just lib' }
Nothing ->
error "Expected a library, but none found"
| otherwise = pd
in f pd' lbi us flags
modHook relocatableBuild f pd lbi us flags
= do let verbosity = normal
idts = updateInstallDirTemplates relocatableBuild
myPrefix myLibdir myDocdir
(installDirTemplates lbi)
progs = withPrograms lbi
stripProgram' = stripProgram {
programFindLocation = \_ _ -> return (Just strip) }
progs' <- configureProgram verbosity stripProgram' progs
let lbi' = lbi {
withPrograms = progs',
installDirTemplates = idts,
withSharedLib = withSharedLibs
}
f pd lbi' us flags
doRegister :: FilePath -> FilePath -> FilePath -> FilePath
-> FilePath -> FilePath -> FilePath -> FilePath -> FilePath
-> String -> [String]
-> IO ()
doRegister directory distDir ghc ghcpkg topdir
myDestDir myPrefix myLibdir myDocdir
relocatableBuildStr args
= withCurrentDirectory directory $ do
relocatableBuild <- case relocatableBuildStr of
"YES" -> return True
"NO" -> return False
_ -> die ["Bad relocatableBuildStr: " ++
show relocatableBuildStr]
let regArgs = "register" : "--builddir" : distDir : args
regHooks = userHooks {
regHook = modHook relocatableBuild
$ regHook userHooks
}
defaultMainWithHooksArgs regHooks regArgs
where
modHook relocatableBuild f pd lbi us flags
= do let verbosity = normal
idts = updateInstallDirTemplates relocatableBuild
myPrefix myLibdir myDocdir
(installDirTemplates lbi)
progs = withPrograms lbi
ghcpkgconf = topdir </> "package.conf.d"
ghcProgram' = ghcProgram {
programPostConf = \_ cp -> return cp { programDefaultArgs = ["-B" ++ topdir] },
programFindLocation = \_ _ -> return (Just ghc) }
ghcPkgProgram' = ghcPkgProgram {
programPostConf = \_ cp -> return cp { programDefaultArgs =
["--global-package-db", ghcpkgconf]
++ ["--force" | not (null myDestDir) ] },
programFindLocation = \_ _ -> return (Just ghcpkg) }
configurePrograms ps conf = foldM (flip (configureProgram verbosity)) conf ps
progs' <- configurePrograms [ghcProgram', ghcPkgProgram'] progs
let Just ghcPkgProg = lookupProgram ghcPkgProgram' progs'
instInfos <- dump verbosity ghcPkgProg GlobalPackageDB
let installedPkgs' = PackageIndex.fromList instInfos
let updateComponentConfig (cn, clbi, deps)
= (cn, updateComponentLocalBuildInfo clbi, deps)
updateComponentLocalBuildInfo clbi
= clbi {
componentPackageDeps =
[ (fixupPackageId instInfos ipid, pid)
| (ipid,pid) <- componentPackageDeps clbi ]
}
ccs' = map updateComponentConfig (componentsConfigs lbi)
lbi' = lbi {
componentsConfigs = ccs',
installedPkgs = installedPkgs',
installDirTemplates = idts,
withPrograms = progs'
}
f pd lbi' us flags
updateInstallDirTemplates :: Bool -> FilePath -> FilePath -> FilePath
-> InstallDirTemplates
-> InstallDirTemplates
updateInstallDirTemplates relocatableBuild myPrefix myLibdir myDocdir idts
= idts {
prefix = toPathTemplate $
if relocatableBuild
then "$topdir"
else myPrefix,
libdir = toPathTemplate $
if relocatableBuild
then "$topdir"
else myLibdir,
libsubdir = toPathTemplate "$pkgid",
docdir = toPathTemplate $
if relocatableBuild
then "$topdir/../doc/html/libraries/$pkgid"
else (myDocdir </> "$pkgid"),
htmldir = toPathTemplate "$docdir"
}
-- The packages are built with the package ID ending in "-inplace", but
-- when they're installed they get the package hash appended. We need to
-- fix up the package deps so that they use the hash package IDs, not
-- the inplace package IDs.
fixupPackageId :: [Installed.InstalledPackageInfo]
-> InstalledPackageId
-> InstalledPackageId
fixupPackageId _ x@(InstalledPackageId ipi)
| "builtin_" `isPrefixOf` ipi = x
fixupPackageId ipinfos (InstalledPackageId ipi)
= case stripPrefix (reverse "-inplace") $ reverse ipi of
Nothing ->
error ("Installed package ID doesn't end in -inplace: " ++ show ipi)
Just x ->
let ipi' = reverse ('-' : x)
f (ipinfo : ipinfos') = case Installed.installedPackageId ipinfo of
y@(InstalledPackageId ipinfoid)
| ipi' `isPrefixOf` ipinfoid ->
y
_ ->
f ipinfos'
f [] = error ("Installed package ID not registered: " ++ show ipi)
in f ipinfos
-- On Windows we need to split the ghc package into 2 pieces, or the
-- DLL that it makes contains too many symbols (#5987). There are
-- therefore 2 libraries, not just the 1 that Cabal assumes.
mangleLbi :: FilePath -> FilePath -> LocalBuildInfo -> LocalBuildInfo
mangleLbi "compiler" "stage2" lbi
| isWindows =
let ccs' = [ (cn, updateComponentLocalBuildInfo clbi, cns)
| (cn, clbi, cns) <- componentsConfigs lbi ]
updateComponentLocalBuildInfo clbi@(LibComponentLocalBuildInfo {})
= let cls' = concat [ [ LibraryName n, LibraryName (n ++ "-0") ]
| LibraryName n <- componentLibraries clbi ]
in clbi { componentLibraries = cls' }
updateComponentLocalBuildInfo clbi = clbi
in lbi { componentsConfigs = ccs' }
where isWindows = case hostPlatform lbi of
Platform _ Windows -> True
_ -> False
mangleLbi _ _ lbi = lbi
generate :: FilePath -> FilePath -> String -> [String] -> IO ()
generate directory distdir dll0Modules config_args
= withCurrentDirectory directory
$ do let verbosity = normal
-- XXX We shouldn't just configure with the default flags
-- XXX And this, and thus the "getPersistBuildConfig distdir" below,
-- aren't going to work when the deps aren't built yet
withArgs (["configure", "--distdir", distdir] ++ config_args)
runDefaultMain
lbi0 <- getPersistBuildConfig distdir
let lbi = mangleLbi directory distdir lbi0
pd0 = localPkgDescr lbi
writePersistBuildConfig distdir lbi
hooked_bi <-
if (buildType pd0 == Just Configure) || (buildType pd0 == Just Custom)
then do
maybe_infoFile <- defaultHookedPackageDesc
case maybe_infoFile of
Nothing -> return emptyHookedBuildInfo
Just infoFile -> readHookedBuildInfo verbosity infoFile
else
return emptyHookedBuildInfo
let pd = updatePackageDescription hooked_bi pd0
-- generate Paths_<pkg>.hs and cabal-macros.h
writeAutogenFiles verbosity pd lbi
-- generate inplace-pkg-config
withLibLBI pd lbi $ \lib clbi ->
do cwd <- getCurrentDirectory
let ipid = InstalledPackageId (display (packageId pd) ++ "-inplace")
let installedPkgInfo = inplaceInstalledPackageInfo cwd distdir
pd lib lbi clbi
final_ipi = installedPkgInfo {
Installed.installedPackageId = ipid,
Installed.haddockHTMLs = []
}
content = Installed.showInstalledPackageInfo final_ipi ++ "\n"
writeFileAtomic (distdir </> "inplace-pkg-config") (BS.pack $ toUTF8 content)
let
libBiModules lib = (libBuildInfo lib, libModules lib)
exeBiModules exe = (buildInfo exe, ModuleName.main : exeModules exe)
biModuless = (maybeToList $ fmap libBiModules $ library pd)
++ (map exeBiModules $ executables pd)
buildableBiModuless = filter isBuildable biModuless
where isBuildable (bi', _) = buildable bi'
(bi, modules) = case buildableBiModuless of
[] -> error "No buildable component found"
[biModules] -> biModules
_ -> error ("XXX ghc-cabal can't handle " ++
"more than one buildinfo yet")
-- XXX Another Just...
Just ghcProg = lookupProgram ghcProgram (withPrograms lbi)
dep_pkgs = PackageIndex.topologicalOrder (packageHacks (installedPkgs lbi))
forDeps f = concatMap f dep_pkgs
-- copied from Distribution.Simple.PreProcess.ppHsc2Hs
packageHacks = case compilerFlavor (compiler lbi) of
GHC -> hackRtsPackage
_ -> id
-- We don't link in the actual Haskell libraries of our
-- dependencies, so the -u flags in the ldOptions of the rts
-- package mean linking fails on OS X (it's ld is a tad
-- stricter than gnu ld). Thus we remove the ldOptions for
-- GHC's rts package:
hackRtsPackage index =
case PackageIndex.lookupPackageName index (PackageName "rts") of
[(_,[rts])] ->
PackageIndex.insert rts{
Installed.ldOptions = [],
Installed.libraryDirs = filter (not . ("gcc-lib" `isSuffixOf`)) (Installed.libraryDirs rts)} index
-- GHC <= 6.12 had $topdir/gcc-lib in their
-- library-dirs for the rts package, which causes
-- problems when we try to use the in-tree mingw,
-- due to accidentally picking up the incompatible
-- libraries there. So we filter out gcc-lib from
-- the RTS's library-dirs here.
_ -> error "No (or multiple) ghc rts package is registered!!"
dep_ids = map snd (externalPackageDeps lbi)
deps = map display dep_ids
depNames = map (display . packageName) dep_ids
transitive_dep_ids = map Installed.sourcePackageId dep_pkgs
transitiveDeps = map display transitive_dep_ids
transitiveDepNames = map (display . packageName) transitive_dep_ids
libraryDirs = forDeps Installed.libraryDirs
-- The mkLibraryRelDir function is a bit of a hack.
-- Ideally it should be handled in the makefiles instead.
mkLibraryRelDir "rts" = "rts/dist/build"
mkLibraryRelDir "ghc" = "compiler/stage2/build"
mkLibraryRelDir "Cabal" = "libraries/Cabal/Cabal/dist-install/build"
mkLibraryRelDir l = "libraries/" ++ l ++ "/dist-install/build"
libraryRelDirs = map mkLibraryRelDir transitiveDepNames
wrappedIncludeDirs <- wrap $ forDeps Installed.includeDirs
wrappedLibraryDirs <- wrap libraryDirs
let variablePrefix = directory ++ '_':distdir
mods = map display modules
otherMods = map display (otherModules bi)
allMods = mods ++ otherMods
let xs = [variablePrefix ++ "_VERSION = " ++ display (pkgVersion (package pd)),
variablePrefix ++ "_MODULES = " ++ unwords mods,
variablePrefix ++ "_HIDDEN_MODULES = " ++ unwords otherMods,
variablePrefix ++ "_SYNOPSIS =" ++ synopsis pd,
variablePrefix ++ "_HS_SRC_DIRS = " ++ unwords (hsSourceDirs bi),
variablePrefix ++ "_DEPS = " ++ unwords deps,
variablePrefix ++ "_DEP_NAMES = " ++ unwords depNames,
variablePrefix ++ "_TRANSITIVE_DEPS = " ++ unwords transitiveDeps,
variablePrefix ++ "_TRANSITIVE_DEP_NAMES = " ++ unwords transitiveDepNames,
variablePrefix ++ "_INCLUDE_DIRS = " ++ unwords (includeDirs bi),
variablePrefix ++ "_INCLUDES = " ++ unwords (includes bi),
variablePrefix ++ "_INSTALL_INCLUDES = " ++ unwords (installIncludes bi),
variablePrefix ++ "_EXTRA_LIBRARIES = " ++ unwords (extraLibs bi),
variablePrefix ++ "_EXTRA_LIBDIRS = " ++ unwords (extraLibDirs bi),
variablePrefix ++ "_C_SRCS = " ++ unwords (cSources bi),
variablePrefix ++ "_CMM_SRCS := $(addprefix cbits/,$(notdir $(wildcard " ++ directory ++ "/cbits/*.cmm)))",
variablePrefix ++ "_DATA_FILES = " ++ unwords (dataFiles pd),
-- XXX This includes things it shouldn't, like:
-- -odir dist-bootstrapping/build
variablePrefix ++ "_HC_OPTS = " ++ escape (unwords
( programDefaultArgs ghcProg
++ hcOptions GHC bi
++ languageToFlags (compiler lbi) (defaultLanguage bi)
++ extensionsToFlags (compiler lbi) (usedExtensions bi)
++ programOverrideArgs ghcProg)),
variablePrefix ++ "_CC_OPTS = " ++ unwords (ccOptions bi),
variablePrefix ++ "_CPP_OPTS = " ++ unwords (cppOptions bi),
variablePrefix ++ "_LD_OPTS = " ++ unwords (ldOptions bi),
variablePrefix ++ "_DEP_INCLUDE_DIRS_SINGLE_QUOTED = " ++ unwords wrappedIncludeDirs,
variablePrefix ++ "_DEP_CC_OPTS = " ++ unwords (forDeps Installed.ccOptions),
variablePrefix ++ "_DEP_LIB_DIRS_SINGLE_QUOTED = " ++ unwords wrappedLibraryDirs,
variablePrefix ++ "_DEP_LIB_DIRS_SEARCHPATH = " ++ mkSearchPath libraryDirs,
variablePrefix ++ "_DEP_LIB_REL_DIRS = " ++ unwords libraryRelDirs,
variablePrefix ++ "_DEP_LIB_REL_DIRS_SEARCHPATH = " ++ mkSearchPath libraryRelDirs,
variablePrefix ++ "_DEP_EXTRA_LIBS = " ++ unwords (forDeps Installed.extraLibraries),
variablePrefix ++ "_DEP_LD_OPTS = " ++ unwords (forDeps Installed.ldOptions),
variablePrefix ++ "_BUILD_GHCI_LIB = " ++ boolToYesNo (withGHCiLib lbi),
"",
-- Sometimes we need to modify the automatically-generated package-data.mk
-- bindings in a special way for the GHC build system, so allow that here:
"$(eval $(" ++ directory ++ "_PACKAGE_MAGIC))"
]
writeFile (distdir ++ "/package-data.mk") $ unlines xs
writeFile (distdir ++ "/haddock-prologue.txt") $
if null (description pd) then synopsis pd
else description pd
unless (null dll0Modules) $
do let dll0Mods = words dll0Modules
dllMods = allMods \\ dll0Mods
dllModSets = map unwords [dll0Mods, dllMods]
writeFile (distdir ++ "/dll-split") $ unlines dllModSets
where
escape = foldr (\c xs -> if c == '#' then '\\':'#':xs else c:xs) []
wrap = mapM wrap1
wrap1 s
| null s = die ["Wrapping empty value"]
| '\'' `elem` s = die ["Single quote in value to be wrapped:", s]
-- We want to be able to assume things like <space><quote> is the
-- start of a value, so check there are no spaces in confusing
-- positions
| head s == ' ' = die ["Leading space in value to be wrapped:", s]
| last s == ' ' = die ["Trailing space in value to be wrapped:", s]
| otherwise = return ("\'" ++ s ++ "\'")
mkSearchPath = intercalate [searchPathSeparator]
boolToYesNo True = "YES"
boolToYesNo False = "NO"
| ekmett/ghc | utils/ghc-cabal/Main.hs | bsd-3-clause | 23,298 | 0 | 23 | 8,374 | 4,641 | 2,366 | 2,275 | 388 | 13 |
-- |
-- Module : $Header$
-- Copyright : (c) 2013-2015 Galois, Inc.
-- License : BSD3
-- Maintainer : [email protected]
-- Stability : provisional
-- Portability : portable
{-# LANGUAGE Safe #-}
{-# LANGUAGE MultiParamTypeClasses, FunctionalDependencies #-}
{-# LANGUAGE UndecidableInstances, FlexibleInstances #-}
{-# LANGUAGE DeriveFunctor #-}
module Cryptol.TypeCheck.TypeMap
( TypeMap(..), TypesMap, TrieMap(..)
, insertTM, insertWithTM
, membersTM
, mapTM, mapWithKeyTM, mapMaybeTM
, List(..)
) where
import Cryptol.TypeCheck.AST
import qualified Data.Map as Map
import Data.Map (Map)
import Data.Maybe(fromMaybe,maybeToList)
import Control.Monad((<=<))
import Data.List(sortBy)
import Data.Maybe (isNothing)
import Data.Ord(comparing)
class TrieMap m k | m -> k where
emptyTM :: m a
nullTM :: m a -> Bool
lookupTM :: k -> m a -> Maybe a
alterTM :: k -> (Maybe a -> Maybe a) -> m a -> m a
unionTM :: (a -> a -> a) -> m a -> m a -> m a
toListTM :: m a -> [(k,a)]
mapMaybeWithKeyTM :: (k -> a -> Maybe b) -> m a -> m b
membersTM :: TrieMap m k => m a -> [a]
membersTM = map snd . toListTM
insertTM :: TrieMap m k => k -> a -> m a -> m a
insertTM t a = alterTM t (\_ -> Just a)
insertWithTM :: TrieMap m k => (a -> a -> a) -> k -> a -> m a -> m a
insertWithTM f t new = alterTM t $ \mb -> Just $ case mb of
Nothing -> new
Just old -> f old new
{-# INLINE mapTM #-}
mapTM :: TrieMap m k => (a -> b) -> m a -> m b
mapTM f = mapMaybeWithKeyTM (\ _ a -> Just (f a))
{-# INLINE mapWithKeyTM #-}
mapWithKeyTM :: TrieMap m k => (k -> a -> b) -> m a -> m b
mapWithKeyTM f = mapMaybeWithKeyTM (\ k a -> Just (f k a))
{-# INLINE mapMaybeTM #-}
mapMaybeTM :: TrieMap m k => (a -> Maybe b) -> m a -> m b
mapMaybeTM f = mapMaybeWithKeyTM (\_ -> f)
data List m a = L { nil :: Maybe a
, cons :: m (List m a)
} deriving (Functor)
instance TrieMap m a => TrieMap (List m) [a] where
emptyTM = L { nil = Nothing, cons = emptyTM }
nullTM k = isNothing (nil k) && nullTM (cons k)
lookupTM k =
case k of
[] -> nil
x : xs -> lookupTM xs <=< lookupTM x . cons
alterTM k f m =
case k of
[] -> m { nil = f (nil m) }
x:xs -> m { cons = alterTM x (updSub xs f) (cons m) }
toListTM m =
[ ([], v) | v <- maybeToList (nil m) ] ++
[ (x:xs,v) | (x,m1) <- toListTM (cons m), (xs,v) <- toListTM m1 ]
unionTM f m1 m2 = L { nil = case (nil m1, nil m2) of
(Just x, Just y) -> Just (f x y)
(Just x, _) -> Just x
(_, Just y) -> Just y
_ -> Nothing
, cons = unionTM (unionTM f) (cons m1) (cons m2)
}
mapMaybeWithKeyTM f = go []
where
go acc l = L { nil = f (reverse acc) =<< nil l
, cons = mapMaybeWithKeyTM (\k a -> Just (go (k:acc) a)) (cons l)
}
instance Ord a => TrieMap (Map a) a where
emptyTM = Map.empty
nullTM = Map.null
lookupTM = Map.lookup
alterTM = flip Map.alter
toListTM = Map.toList
unionTM = Map.unionWith
mapMaybeWithKeyTM = Map.mapMaybeWithKey
type TypesMap = List TypeMap
data TypeMap a = TM { tvar :: Map TVar a
, tcon :: Map TCon (List TypeMap a)
, trec :: Map [Name] (List TypeMap a)
} deriving (Functor)
instance TrieMap TypeMap Type where
emptyTM = TM { tvar = emptyTM, tcon = emptyTM, trec = emptyTM }
nullTM ty = and [ nullTM (tvar ty)
, nullTM (tcon ty)
, nullTM (trec ty) ]
lookupTM ty =
case ty of
TUser _ _ t -> lookupTM t
TVar x -> lookupTM x . tvar
TCon c ts -> lookupTM ts <=< lookupTM c . tcon
TRec fs -> let (xs,ts) = unzip $ sortBy (comparing fst) fs
in lookupTM ts <=< lookupTM xs . trec
alterTM ty f m =
case ty of
TUser _ _ t -> alterTM t f m
TVar x -> m { tvar = alterTM x f (tvar m) }
TCon c ts -> m { tcon = alterTM c (updSub ts f) (tcon m) }
TRec fs -> let (xs,ts) = unzip $ sortBy (comparing fst) fs
in m { trec = alterTM xs (updSub ts f) (trec m) }
toListTM m =
[ (TVar x, v) | (x,v) <- toListTM (tvar m) ] ++
[ (TCon c ts, v) | (c,m1) <- toListTM (tcon m)
, (ts,v) <- toListTM m1 ] ++
[ (TRec (zip fs ts), v) | (fs,m1) <- toListTM (trec m)
, (ts,v) <- toListTM m1 ]
unionTM f m1 m2 = TM { tvar = unionTM f (tvar m1) (tvar m2)
, tcon = unionTM (unionTM f) (tcon m1) (tcon m2)
, trec = unionTM (unionTM f) (trec m1) (trec m2)
}
mapMaybeWithKeyTM f m =
TM { tvar = mapMaybeWithKeyTM (\v -> f (TVar v)) (tvar m)
, tcon = mapWithKeyTM (\c l -> mapMaybeWithKeyTM
(\ts a -> f (TCon c ts) a) l) (tcon m)
, trec = mapWithKeyTM (\fs l -> mapMaybeWithKeyTM
(\ts a -> f (TRec (zip fs ts)) a) l) (trec m)
}
updSub :: TrieMap m k => k -> (Maybe a -> Maybe a) -> Maybe (m a) -> Maybe (m a)
updSub k f = Just . alterTM k f . fromMaybe emptyTM
instance Show a => Show (TypeMap a) where
showsPrec p xs = showsPrec p (toListTM xs)
| iblumenfeld/cryptol | src/Cryptol/TypeCheck/TypeMap.hs | bsd-3-clause | 5,626 | 0 | 18 | 2,079 | 2,373 | 1,229 | 1,144 | -1 | -1 |
<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE helpset PUBLIC "-//Sun Microsystems Inc.//DTD JavaHelp HelpSet Version 2.0//EN" "http://java.sun.com/products/javahelp/helpset_2_0.dtd">
<helpset version="2.0" xml:lang="fa-IR">
<title>Custom Payloads Add-on</title>
<maps>
<homeID>custompayloads</homeID>
<mapref location="map.jhm"/>
</maps>
<view>
<name>TOC</name>
<label>Contents</label>
<type>org.zaproxy.zap.extension.help.ZapTocView</type>
<data>toc.xml</data>
</view>
<view>
<name>Index</name>
<label>Index</label>
<type>javax.help.IndexView</type>
<data>index.xml</data>
</view>
<view>
<name>Search</name>
<label>Search</label>
<type>javax.help.SearchView</type>
<data engine="com.sun.java.help.search.DefaultSearchEngine">
JavaHelpSearch
</data>
</view>
<view>
<name>Favorites</name>
<label>Favorites</label>
<type>javax.help.FavoritesView</type>
</view>
</helpset> | thc202/zap-extensions | addOns/custompayloads/src/main/javahelp/org/zaproxy/zap/extension/custompayloads/resources/help_fa_IR/helpset_fa_IR.hs | apache-2.0 | 978 | 82 | 53 | 157 | 396 | 209 | 187 | -1 | -1 |
{-# LANGUAGE CPP #-}
module CmmType
( CmmType -- Abstract
, b8, b16, b32, b64, b128, b256, b512, f32, f64, bWord, bHalfWord, gcWord
, cInt, cLong
, cmmBits, cmmFloat
, typeWidth, cmmEqType, cmmEqType_ignoring_ptrhood
, isFloatType, isGcPtrType, isWord32, isWord64, isFloat64, isFloat32
, Width(..)
, widthInBits, widthInBytes, widthInLog, widthFromBytes
, wordWidth, halfWordWidth, cIntWidth, cLongWidth
, halfWordMask
, narrowU, narrowS
, rEP_CostCentreStack_mem_alloc
, rEP_CostCentreStack_scc_count
, rEP_StgEntCounter_allocs
, rEP_StgEntCounter_allocd
, ForeignHint(..)
, Length
, vec, vec2, vec4, vec8, vec16
, vec2f64, vec2b64, vec4f32, vec4b32, vec8b16, vec16b8
, cmmVec
, vecLength, vecElemType
, isVecType
)
where
#include "HsVersions.h"
import DynFlags
import FastString
import Outputable
import Data.Word
import Data.Int
-----------------------------------------------------------------------------
-- CmmType
-----------------------------------------------------------------------------
-- NOTE: CmmType is an abstract type, not exported from this
-- module so you can easily change its representation
--
-- However Width is exported in a concrete way,
-- and is used extensively in pattern-matching
data CmmType -- The important one!
= CmmType CmmCat Width
data CmmCat -- "Category" (not exported)
= GcPtrCat -- GC pointer
| BitsCat -- Non-pointer
| FloatCat -- Float
| VecCat Length CmmCat -- Vector
deriving( Eq )
-- See Note [Signed vs unsigned] at the end
instance Outputable CmmType where
ppr (CmmType cat wid) = ppr cat <> ppr (widthInBits wid)
instance Outputable CmmCat where
ppr FloatCat = text "F"
ppr GcPtrCat = text "P"
ppr BitsCat = text "I"
ppr (VecCat n cat) = ppr cat <> text "x" <> ppr n <> text "V"
-- Why is CmmType stratified? For native code generation,
-- most of the time you just want to know what sort of register
-- to put the thing in, and for this you need to know how
-- many bits thing has and whether it goes in a floating-point
-- register. By contrast, the distinction between GcPtr and
-- GcNonPtr is of interest to only a few parts of the code generator.
-------- Equality on CmmType --------------
-- CmmType is *not* an instance of Eq; sometimes we care about the
-- Gc/NonGc distinction, and sometimes we don't
-- So we use an explicit function to force you to think about it
cmmEqType :: CmmType -> CmmType -> Bool -- Exact equality
cmmEqType (CmmType c1 w1) (CmmType c2 w2) = c1==c2 && w1==w2
cmmEqType_ignoring_ptrhood :: CmmType -> CmmType -> Bool
-- This equality is temporary; used in CmmLint
-- but the RTS files are not yet well-typed wrt pointers
cmmEqType_ignoring_ptrhood (CmmType c1 w1) (CmmType c2 w2)
= c1 `weak_eq` c2 && w1==w2
where
weak_eq :: CmmCat -> CmmCat -> Bool
FloatCat `weak_eq` FloatCat = True
FloatCat `weak_eq` _other = False
_other `weak_eq` FloatCat = False
(VecCat l1 cat1) `weak_eq` (VecCat l2 cat2) = l1 == l2
&& cat1 `weak_eq` cat2
(VecCat {}) `weak_eq` _other = False
_other `weak_eq` (VecCat {}) = False
_word1 `weak_eq` _word2 = True -- Ignores GcPtr
--- Simple operations on CmmType -----
typeWidth :: CmmType -> Width
typeWidth (CmmType _ w) = w
cmmBits, cmmFloat :: Width -> CmmType
cmmBits = CmmType BitsCat
cmmFloat = CmmType FloatCat
-------- Common CmmTypes ------------
-- Floats and words of specific widths
b8, b16, b32, b64, b128, b256, b512, f32, f64 :: CmmType
b8 = cmmBits W8
b16 = cmmBits W16
b32 = cmmBits W32
b64 = cmmBits W64
b128 = cmmBits W128
b256 = cmmBits W256
b512 = cmmBits W512
f32 = cmmFloat W32
f64 = cmmFloat W64
-- CmmTypes of native word widths
bWord :: DynFlags -> CmmType
bWord dflags = cmmBits (wordWidth dflags)
bHalfWord :: DynFlags -> CmmType
bHalfWord dflags = cmmBits (halfWordWidth dflags)
gcWord :: DynFlags -> CmmType
gcWord dflags = CmmType GcPtrCat (wordWidth dflags)
cInt, cLong :: DynFlags -> CmmType
cInt dflags = cmmBits (cIntWidth dflags)
cLong dflags = cmmBits (cLongWidth dflags)
------------ Predicates ----------------
isFloatType, isGcPtrType :: CmmType -> Bool
isFloatType (CmmType FloatCat _) = True
isFloatType _other = False
isGcPtrType (CmmType GcPtrCat _) = True
isGcPtrType _other = False
isWord32, isWord64, isFloat32, isFloat64 :: CmmType -> Bool
-- isWord64 is true of 64-bit non-floats (both gc-ptrs and otherwise)
-- isFloat32 and 64 are obvious
isWord64 (CmmType BitsCat W64) = True
isWord64 (CmmType GcPtrCat W64) = True
isWord64 _other = False
isWord32 (CmmType BitsCat W32) = True
isWord32 (CmmType GcPtrCat W32) = True
isWord32 _other = False
isFloat32 (CmmType FloatCat W32) = True
isFloat32 _other = False
isFloat64 (CmmType FloatCat W64) = True
isFloat64 _other = False
-----------------------------------------------------------------------------
-- Width
-----------------------------------------------------------------------------
data Width = W8 | W16 | W32 | W64
| W80 -- Extended double-precision float,
-- used in x86 native codegen only.
-- (we use Ord, so it'd better be in this order)
| W128
| W256
| W512
deriving (Eq, Ord, Show)
instance Outputable Width where
ppr rep = ptext (mrStr rep)
mrStr :: Width -> LitString
mrStr W8 = sLit("W8")
mrStr W16 = sLit("W16")
mrStr W32 = sLit("W32")
mrStr W64 = sLit("W64")
mrStr W128 = sLit("W128")
mrStr W256 = sLit("W256")
mrStr W512 = sLit("W512")
mrStr W80 = sLit("W80")
-------- Common Widths ------------
wordWidth :: DynFlags -> Width
wordWidth dflags
| wORD_SIZE dflags == 4 = W32
| wORD_SIZE dflags == 8 = W64
| otherwise = panic "MachOp.wordRep: Unknown word size"
halfWordWidth :: DynFlags -> Width
halfWordWidth dflags
| wORD_SIZE dflags == 4 = W16
| wORD_SIZE dflags == 8 = W32
| otherwise = panic "MachOp.halfWordRep: Unknown word size"
halfWordMask :: DynFlags -> Integer
halfWordMask dflags
| wORD_SIZE dflags == 4 = 0xFFFF
| wORD_SIZE dflags == 8 = 0xFFFFFFFF
| otherwise = panic "MachOp.halfWordMask: Unknown word size"
-- cIntRep is the Width for a C-language 'int'
cIntWidth, cLongWidth :: DynFlags -> Width
cIntWidth dflags = case cINT_SIZE dflags of
4 -> W32
8 -> W64
s -> panic ("cIntWidth: Unknown cINT_SIZE: " ++ show s)
cLongWidth dflags = case cLONG_SIZE dflags of
4 -> W32
8 -> W64
s -> panic ("cIntWidth: Unknown cLONG_SIZE: " ++ show s)
widthInBits :: Width -> Int
widthInBits W8 = 8
widthInBits W16 = 16
widthInBits W32 = 32
widthInBits W64 = 64
widthInBits W128 = 128
widthInBits W256 = 256
widthInBits W512 = 512
widthInBits W80 = 80
widthInBytes :: Width -> Int
widthInBytes W8 = 1
widthInBytes W16 = 2
widthInBytes W32 = 4
widthInBytes W64 = 8
widthInBytes W128 = 16
widthInBytes W256 = 32
widthInBytes W512 = 64
widthInBytes W80 = 10
widthFromBytes :: Int -> Width
widthFromBytes 1 = W8
widthFromBytes 2 = W16
widthFromBytes 4 = W32
widthFromBytes 8 = W64
widthFromBytes 16 = W128
widthFromBytes 32 = W256
widthFromBytes 64 = W512
widthFromBytes 10 = W80
widthFromBytes n = pprPanic "no width for given number of bytes" (ppr n)
-- log_2 of the width in bytes, useful for generating shifts.
widthInLog :: Width -> Int
widthInLog W8 = 0
widthInLog W16 = 1
widthInLog W32 = 2
widthInLog W64 = 3
widthInLog W128 = 4
widthInLog W256 = 5
widthInLog W512 = 6
widthInLog W80 = panic "widthInLog: F80"
-- widening / narrowing
narrowU :: Width -> Integer -> Integer
narrowU W8 x = fromIntegral (fromIntegral x :: Word8)
narrowU W16 x = fromIntegral (fromIntegral x :: Word16)
narrowU W32 x = fromIntegral (fromIntegral x :: Word32)
narrowU W64 x = fromIntegral (fromIntegral x :: Word64)
narrowU _ _ = panic "narrowTo"
narrowS :: Width -> Integer -> Integer
narrowS W8 x = fromIntegral (fromIntegral x :: Int8)
narrowS W16 x = fromIntegral (fromIntegral x :: Int16)
narrowS W32 x = fromIntegral (fromIntegral x :: Int32)
narrowS W64 x = fromIntegral (fromIntegral x :: Int64)
narrowS _ _ = panic "narrowTo"
-----------------------------------------------------------------------------
-- SIMD
-----------------------------------------------------------------------------
type Length = Int
vec :: Length -> CmmType -> CmmType
vec l (CmmType cat w) = CmmType (VecCat l cat) vecw
where
vecw :: Width
vecw = widthFromBytes (l*widthInBytes w)
vec2, vec4, vec8, vec16 :: CmmType -> CmmType
vec2 = vec 2
vec4 = vec 4
vec8 = vec 8
vec16 = vec 16
vec2f64, vec2b64, vec4f32, vec4b32, vec8b16, vec16b8 :: CmmType
vec2f64 = vec 2 f64
vec2b64 = vec 2 b64
vec4f32 = vec 4 f32
vec4b32 = vec 4 b32
vec8b16 = vec 8 b16
vec16b8 = vec 16 b8
cmmVec :: Int -> CmmType -> CmmType
cmmVec n (CmmType cat w) =
CmmType (VecCat n cat) (widthFromBytes (n*widthInBytes w))
vecLength :: CmmType -> Length
vecLength (CmmType (VecCat l _) _) = l
vecLength _ = panic "vecLength: not a vector"
vecElemType :: CmmType -> CmmType
vecElemType (CmmType (VecCat l cat) w) = CmmType cat scalw
where
scalw :: Width
scalw = widthFromBytes (widthInBytes w `div` l)
vecElemType _ = panic "vecElemType: not a vector"
isVecType :: CmmType -> Bool
isVecType (CmmType (VecCat {}) _) = True
isVecType _ = False
-------------------------------------------------------------------------
-- Hints
-- Hints are extra type information we attach to the arguments and
-- results of a foreign call, where more type information is sometimes
-- needed by the ABI to make the correct kind of call.
data ForeignHint
= NoHint | AddrHint | SignedHint
deriving( Eq )
-- Used to give extra per-argument or per-result
-- information needed by foreign calling conventions
-------------------------------------------------------------------------
-- These don't really belong here, but I don't know where is best to
-- put them.
rEP_CostCentreStack_mem_alloc :: DynFlags -> CmmType
rEP_CostCentreStack_mem_alloc dflags
= cmmBits (widthFromBytes (pc_REP_CostCentreStack_mem_alloc pc))
where pc = sPlatformConstants (settings dflags)
rEP_CostCentreStack_scc_count :: DynFlags -> CmmType
rEP_CostCentreStack_scc_count dflags
= cmmBits (widthFromBytes (pc_REP_CostCentreStack_scc_count pc))
where pc = sPlatformConstants (settings dflags)
rEP_StgEntCounter_allocs :: DynFlags -> CmmType
rEP_StgEntCounter_allocs dflags
= cmmBits (widthFromBytes (pc_REP_StgEntCounter_allocs pc))
where pc = sPlatformConstants (settings dflags)
rEP_StgEntCounter_allocd :: DynFlags -> CmmType
rEP_StgEntCounter_allocd dflags
= cmmBits (widthFromBytes (pc_REP_StgEntCounter_allocd pc))
where pc = sPlatformConstants (settings dflags)
-------------------------------------------------------------------------
{- Note [Signed vs unsigned]
~~~~~~~~~~~~~~~~~~~~~~~~~
Should a CmmType include a signed vs. unsigned distinction?
This is very much like a "hint" in C-- terminology: it isn't necessary
in order to generate correct code, but it might be useful in that the
compiler can generate better code if it has access to higher-level
hints about data. This is important at call boundaries, because the
definition of a function is not visible at all of its call sites, so
the compiler cannot infer the hints.
Here in Cmm, we're taking a slightly different approach. We include
the int vs. float hint in the CmmType, because (a) the majority of
platforms have a strong distinction between float and int registers,
and (b) we don't want to do any heavyweight hint-inference in the
native code backend in order to get good code. We're treating the
hint more like a type: our Cmm is always completely consistent with
respect to hints. All coercions between float and int are explicit.
What about the signed vs. unsigned hint? This information might be
useful if we want to keep sub-word-sized values in word-size
registers, which we must do if we only have word-sized registers.
On such a system, there are two straightforward conventions for
representing sub-word-sized values:
(a) Leave the upper bits undefined. Comparison operations must
sign- or zero-extend both operands before comparing them,
depending on whether the comparison is signed or unsigned.
(b) Always keep the values sign- or zero-extended as appropriate.
Arithmetic operations must narrow the result to the appropriate
size.
A clever compiler might not use either (a) or (b) exclusively, instead
it would attempt to minimize the coercions by analysis: the same kind
of analysis that propagates hints around. In Cmm we don't want to
have to do this, so we plump for having richer types and keeping the
type information consistent.
If signed/unsigned hints are missing from CmmType, then the only
choice we have is (a), because we don't know whether the result of an
operation should be sign- or zero-extended.
Many architectures have extending load operations, which work well
with (b). To make use of them with (a), you need to know whether the
value is going to be sign- or zero-extended by an enclosing comparison
(for example), which involves knowing above the context. This is
doable but more complex.
Further complicating the issue is foreign calls: a foreign calling
convention can specify that signed 8-bit quantities are passed as
sign-extended 32 bit quantities, for example (this is the case on the
PowerPC). So we *do* need sign information on foreign call arguments.
Pros for adding signed vs. unsigned to CmmType:
- It would let us use convention (b) above, and get easier
code generation for extending loads.
- Less information required on foreign calls.
- MachOp type would be simpler
Cons:
- More complexity
- What is the CmmType for a VanillaReg? Currently it is
always wordRep, but now we have to decide whether it is
signed or unsigned. The same VanillaReg can thus have
different CmmType in different parts of the program.
- Extra coercions cluttering up expressions.
Currently for GHC, the foreign call point is moot, because we do our
own promotion of sub-word-sized values to word-sized values. The Int8
type is represented by an Int# which is kept sign-extended at all times
(this is slightly naughty, because we're making assumptions about the
C calling convention rather early on in the compiler). However, given
this, the cons outweigh the pros.
-}
| oldmanmike/ghc | compiler/cmm/CmmType.hs | bsd-3-clause | 15,086 | 0 | 11 | 3,373 | 2,796 | 1,498 | 1,298 | 238 | 7 |
module RecordIn4 where
data S = S1 { x :: Int } | S2 { x :: Int } deriving Show
{- map2 xs = map (\y -> y {x = 1}) xs -}
map2 xs = (case ((\ y -> y {x = 1}), xs) of
(f, []) -> []
(f, (x : xs)) -> (f x) : (map f xs))
| SAdams601/HaRe | old/testing/generativeFold/RecordIn4.hs | bsd-3-clause | 251 | 0 | 11 | 97 | 124 | 73 | 51 | 5 | 2 |
module Renaming.C1 where
import Renaming.D1
instance SameOrNot Double where
isSame a b = a ==b
isNotSame a b = a /=b
myFringe:: Tree a -> [a]
myFringe (Leaf x ) = [x]
myFringe (Branch left right) = myFringe left
| mpickering/HaRe | test/testdata/Renaming/C1.hs | bsd-3-clause | 229 | 0 | 7 | 56 | 98 | 51 | 47 | 8 | 1 |
{-| Scans page of Markdown looking for http links. When it finds them, it submits them
to webcitation.org / https://secure.wikimedia.org/wikipedia/en/wiki/WebCite
(It will also submit them to Alexa (the source for the Internet Archive), but Alexa says that
its bots take weeks to visit and may not ever.)
This module employs the archiver daemon <http://hackage.haskell.org/package/archiver> as a library; `cabal install archiver` will install it.
Limitations:
* Only parses Markdown, not ReST or any other format; this is because 'readMarkdown'
is hardwired into it.
* No rate limitation or choking; will fire off all requests as fast as possible.
If pages have more than 20 external links or so, this may result in your IP being temporarily
banned by WebCite. To avoid this, you can use WebArchiverBot.hs instead, which will parse & dump
URLs into a file processed by the archiver daemon (which *is* rate-limited).
By: Gwern Branwen; placed in the public domain -}
module WebArchiver (plugin) where
import Control.Concurrent (forkIO)
import Network.URL.Archiver as A (checkArchive)
import Network.Gitit.Interface (askUser, bottomUpM, liftIO, uEmail, Plugin(PreCommitTransform), Inline(Link))
import Text.Pandoc (defaultParserState, readMarkdown)
plugin :: Plugin
plugin = PreCommitTransform archivePage
-- archivePage :: String -> ReaderT PluginData (StateT Context IO) String
archivePage x = do mbUser <- askUser
let email = case mbUser of
Nothing -> "[email protected]"
Just u -> uEmail u
let p = readMarkdown defaultParserState x
-- force evaluation and archiving side-effects
_p' <- liftIO $ bottomUpM (archiveLinks email) p
return x -- note: this is read-only - don't actually change page!
archiveLinks :: String -> Inline -> IO Inline
archiveLinks e x@(Link _ (uln, _)) = forkIO (A.checkArchive e uln) >> return x
archiveLinks _ x = return x
| imuli/gitit | plugins/WebArchiver.hs | gpl-2.0 | 2,029 | 0 | 13 | 451 | 260 | 140 | 120 | 17 | 2 |
module Graphics.UI.Bottle.Widgets.Edges(
makeVertical
) where
import Control.Lens ((^.))
import Control.Monad (mplus)
import Data.List (minimumBy)
import Data.Monoid (Monoid(..))
import Data.Ord (comparing)
import Data.Vector.Vector2 (Vector2(..))
import Graphics.UI.Bottle.Direction (Direction)
import Graphics.UI.Bottle.Rect (Rect(..))
import Graphics.UI.Bottle.Widget (Widget(..))
import Graphics.UI.Bottle.Widgets.StdKeys (DirKeys(..), stdDirKeys)
import qualified Control.Lens as Lens
import qualified Graphics.UI.Bottle.Direction as Direction
import qualified Graphics.UI.Bottle.EventMap as EventMap
import qualified Graphics.UI.Bottle.Rect as Rect
import qualified Graphics.UI.Bottle.Widget as Widget
choose ::
Widget.EnterResult f -> Widget.EnterResult f ->
Direction -> Widget.EnterResult f
choose x _ Direction.Outside = x
choose x y (Direction.PrevFocalArea rect) = chooseRect x y rect
choose x y (Direction.Point pt) = chooseRect x y $ Rect pt 0
chooseRect :: Widget.EnterResult f -> Widget.EnterResult f -> Rect -> Widget.EnterResult f
chooseRect x y rect =
minimumOn (Rect.distance rect . (^. Widget.enterResultRect)) [x, y]
where
minimumOn = minimumBy . comparing
makeVertical :: Widget.Size -> Widget f -> Widget f -> Widget f
makeVertical size top unTranslatedBottom = Widget
{ _wIsFocused = _wIsFocused top || _wIsFocused bottom
, _wSize = size
, _wFrame = _wFrame top `mappend` _wFrame bottom
, _wMaybeEnter = mEnter (_wMaybeEnter top) (_wMaybeEnter bottom)
, _wEventMap = eventMap
, _wFocalArea = maybe (Rect 0 0) _wFocalArea selectedWidget
}
where
mEnter (Just enterTop) (Just enterBottom) =
Just $ \dir -> choose (enterTop dir) (enterBottom dir) dir
mEnter x y = x `mplus` y
selectedWidget
| _wIsFocused top = Just $ addTo (EventMap.Doc ["Navigation", "Move", "down"]) (keysDown stdDirKeys) bottom top
| _wIsFocused bottom = Just $ addTo (EventMap.Doc ["Navigation", "Move", "up"]) (keysUp stdDirKeys) top bottom
| otherwise = Nothing
mkKeys = map $ EventMap.ModKey EventMap.noMods
eventMap = maybe mempty _wEventMap selectedWidget
addTo doc ks other me =
maybe id
(Widget.weakerEvents . mkEventMap me doc (mkKeys ks))
(_wMaybeEnter other) me
mkEventMap me doc keys enterOther =
EventMap.keyPresses keys doc . (^. Widget.enterResultEvent) . enterOther .
Direction.PrevFocalArea $ _wFocalArea me
bottom = Widget.translate (Vector2 0 (max topHeight bottomsTop)) unTranslatedBottom
topHeight = _wSize top ^. Lens._2
bottomHeight = _wSize unTranslatedBottom ^. Lens._2
bottomsTop = size ^. Lens._2 - bottomHeight
| sinelaw/lamdu | bottlelib/Graphics/UI/Bottle/Widgets/Edges.hs | gpl-3.0 | 2,665 | 0 | 12 | 463 | 893 | 487 | 406 | 55 | 2 |
{-# LANGUAGE ForeignFunctionInterface, CPP #-}
-- Test the LANGUAGE pragma
module ShouldCompile where
#if 1
foreign import ccall "foo" foo :: Int -> IO Int
#endif
| wxwxwwxxx/ghc | testsuite/tests/parser/should_compile/read039.hs | bsd-3-clause | 164 | 0 | 7 | 27 | 27 | 17 | 10 | 3 | 0 |
{-# LANGUAGE MultiParamTypeClasses, FlexibleInstances #-}
-- !!! One method class from Sergey Mechveliani
-- showed up problematic newtype dict rep.
module Main where
import Data.Ratio
class MBConvertible a b where cm :: a -> b -> Maybe b
c :: MBConvertible a b => a -> b -> b
c a b = case cm a b
of
Just b' -> b'
_ -> error "c a b failed"
instance MBConvertible Int Int where cm a _ = Just a
instance (MBConvertible a b,Integral b) => MBConvertible a (Ratio b)
where
cm a f = case cm a (numerator f) of Just a' -> Just (a'%1)
_ -> Nothing
main = let f = 1%1 :: Ratio Int
n2 = 2::Int
g = (c n2 f) + f
in
putStr (shows g "\n")
| urbanslug/ghc | testsuite/tests/typecheck/should_run/tcrun003.hs | bsd-3-clause | 902 | 0 | 11 | 409 | 266 | 134 | 132 | 16 | 2 |
{-|
Module: Flaw.UI.DefaultStyle.Data
Description: Embedded data for default style.
License: MIT
-}
{-# LANGUAGE TemplateHaskell #-}
module Flaw.UI.DefaultStyle.Data
( loadFontData
) where
import qualified Data.ByteString as B
import qualified Data.ByteString.Lazy as BL
import Flaw.Build
loadFontData :: IO B.ByteString
loadFontData = $(embedIOExp =<< BL.toStrict <$> loadFile "src/DejaVuSans.ttf")
| quyse/flaw | flaw-ui-default-style-data/Flaw/UI/DefaultStyle/Data.hs | mit | 409 | 0 | 9 | 55 | 69 | 43 | 26 | 8 | 1 |
{-# htermination (fromIntegral :: MyInt -> Float) #-}
import qualified Prelude
data MyBool = MyTrue | MyFalse
data List a = Cons a (List a) | Nil
data Float = Float MyInt MyInt ;
data Integer = Integer MyInt ;
data MyInt = Pos Nat | Neg Nat ;
data Nat = Succ Nat | Zero ;
primIntToFloat :: MyInt -> Float;
primIntToFloat x = Float x (Pos (Succ Zero));
primIntegerToFloat :: Integer -> Float;
primIntegerToFloat (Integer x) = primIntToFloat x;
fromIntegerFloat :: Integer -> Float
fromIntegerFloat = primIntegerToFloat;
pt :: (c -> b) -> (a -> c) -> a -> b;
pt f g x = f (g x);
toIntegerMyInt :: MyInt -> Integer
toIntegerMyInt x = Integer x;
fromIntegral = pt fromIntegerFloat toIntegerMyInt;
| ComputationWithBoundedResources/ara-inference | doc/tpdb_trs/Haskell/basic_haskell/fromIntegral_2.hs | mit | 729 | 0 | 9 | 161 | 252 | 141 | 111 | 18 | 1 |
{-# LANGUAGE TemplateHaskell
, TypeFamilies
, OverloadedStrings #-}
module FirewallModel (
Rule(..)
, View(..)
) where
import Generics.BiGUL.TH
import GHC.Generics
import Data.Aeson
data View = View {
rules :: [Rule]
} deriving (Show, Eq)
data Rule = Rule {
ruleID :: String
, securityGroupRefFrom :: String
, securityGroupRefTo :: String
, port :: String
, protocol :: String
} deriving (Show, Eq)
instance FromJSON View where
parseJSON (Object v) = View <$>
v .: "rules"
-- A non-Object value is of the wrong type, so fail.
parseJSON _ = mempty
instance ToJSON View where
-- this generates a Value
toJSON (View rules) =
object ["rules" .= rules]
instance FromJSON Rule where
parseJSON (Object v) = Rule <$>
v .: "ruleID" <*>
v .: "securityGroupRefFrom" <*>
v .: "securityGroupRefTo" <*>
v .: "port" <*>
v .: "protocol"
-- A non-Object value is of the wrong type, so fail.
parseJSON _ = mempty
instance ToJSON Rule where
-- this generates a Value
toJSON (Rule ruleID securityGroupRefFrom securityGroupRefTo port protocol) =
object ["ruleID" .= ruleID
, "securityGroupRefFrom" .= securityGroupRefFrom
, "securityGroupRefTo" .= securityGroupRefTo
, "port" .= port
, "protocol" .= protocol]
deriveBiGULGeneric ''View
deriveBiGULGeneric ''Rule
| prl-tokyo/MAPE-knowledge-base | Haskell/views/FirewallModel.hs | mit | 1,572 | 0 | 15 | 523 | 344 | 189 | 155 | -1 | -1 |
-- Caffeine Script
-- http://www.codewars.com/kata/5434283682b0fdb0420000e6
module Codewars.Kata.Caffeine where
caffeineBuzz :: Integer -> String
caffeineBuzz n | n `mod` 12 == 0 = "CoffeeScript"
| n `mod` 6 == 0 = "JavaScript"
| n `mod` 3 == 0 = "Java"
| otherwise = "mocha_missing!"
| gafiatulin/codewars | src/7 kyu/Caffeine.hs | mit | 334 | 0 | 9 | 92 | 89 | 48 | 41 | 6 | 1 |
{-# LANGUAGE DeriveFunctor #-}
{-# LANGUAGE ViewPatterns #-}
module Snipcheck where
import Control.Monad
import Control.Exception
import Control.Monad.IO.Class
import Data.Char (isSpace)
import Data.List (dropWhileEnd)
import Data.Maybe
import System.Process(readCreateProcess, shell)
import Text.Pandoc (Block(..))
import qualified Data.Text.IO as Text
import qualified Data.Map as Map
import qualified Text.Pandoc as Pandoc
data Sloppy a = Skip | Must a deriving (Show, Functor)
sloppyString :: String -> Sloppy String
sloppyString "..." = Skip
sloppyString str = Must str
checkSloppy :: Eq a => [a] -> [Sloppy a] -> Bool
checkSloppy (a:as) (Must a':as')
| a == a' = checkSloppy as as'
| otherwise = False
checkSloppy (a:as) as'@(Skip:Must a':as'')
| a == a' = checkSloppy as as''
| otherwise = checkSloppy as as'
checkSloppy as (Skip:Skip:as') = checkSloppy as (Skip:as')
checkSloppy [] (Must{}:_) = False
checkSloppy [] (Skip:as') = checkSloppy [] as'
checkSloppy [] [] = True
checkSloppy (_:_) [] = False
checkSloppy _ [Skip] = True
checkMarkdownFile :: FilePath -> IO ()
checkMarkdownFile fp = do
content <- Text.readFile fp
eres <- Pandoc.runIO $ do
Pandoc.Pandoc meta blocks <- Pandoc.readMarkdown Pandoc.def content
let
sections = findSections meta
blocks' =
if null sections
then blocks
else filterBlocksBySectionName sections blocks
forM_ blocks' check
case eres of
Right () -> pure ()
Left e -> throwIO $ userError $ show e
data AcceptSection
= GoodSection
| BadSection
| Dunno
filterBlocksBySectionName :: [String] -> [Pandoc.Block] -> [Pandoc.Block]
filterBlocksBySectionName secs = skipThese
where
skipThese, keepThese :: [Pandoc.Block] -> [Pandoc.Block]
skipThese (b:bs) =
case acceptSection b of
GoodSection -> keepThese bs
_ -> skipThese bs
skipThese [] = []
keepThese (b:bs) = b : case acceptSection b of
BadSection -> skipThese bs
_ -> keepThese bs
keepThese [] = []
acceptSection :: Pandoc.Block -> AcceptSection
acceptSection (Pandoc.Header _ (hName,_,_) _)
| hName `elem` secs = GoodSection
| otherwise = BadSection
acceptSection _ = Dunno
findSections :: Pandoc.Meta -> [String]
findSections (Pandoc.unMeta -> meta) =
case Map.lookup "sc_check-sections" meta of
Just (Pandoc.MetaList ss) -> join $ unMetaString <$> ss
_ -> []
where
unMetaString :: Pandoc.MetaValue -> [String]
unMetaString (Pandoc.MetaString s) =[s]
unMetaString (Pandoc.MetaInlines is) = mapMaybe unMetaStr is
unMetaString _ = []
unMetaStr :: Pandoc.Inline -> Maybe String
unMetaStr (Pandoc.Str s) = Just s
unMetaStr _ = Nothing
trim :: String -> String
trim = dropWhile isSpace . dropWhileEnd isSpace
check :: MonadIO m => Pandoc.Block -> m ()
check (CodeBlock (typ, classes, kvs) content)
| "shell" `elem` classes = do
let Right cmds = extractCommands content
forM_ cmds $ \(cmd, expected) -> do
actual <- (fmap trim . lines) <$> liftIO (readCreateProcess (shell cmd) "")
let expected' = (sloppyString . trim) <$> expected
unless (checkSloppy actual expected') $ error $ mconcat
[ "Couldnt match expected ", show expected'
, " with " <> show actual
]
| otherwise = liftIO $ print (typ, classes, kvs)
check _ = return ()
extractCommands :: String -> Either String [(String, [String])]
extractCommands str = go (lines str)
where
go :: [String] -> Either String [(String, [String])]
go (l:ls) | Just cmd <- toCommand l =
let (output, rest) = break isCommand ls
in ((cmd,output):) <$> go rest
| otherwise = Left $ "Expected a command, got " <> l
go [] = Right []
toCommand :: String -> Maybe String
toCommand ('$':cmd) = Just cmd
toCommand _ = Nothing
isCommand :: String -> Bool
isCommand = isJust . toCommand
someFunc :: IO ()
someFunc = putStrLn "someFunc"
| nmattia/snipcheck | src/Snipcheck.hs | mit | 4,018 | 0 | 18 | 932 | 1,521 | 779 | 742 | 108 | 6 |
import Avus.Scan
import Criterion.Types
import Criterion.Main
noopProcessData :: FilePath -> IO ()
noopProcessData fp = processData (Just fp) (Just "null") $
processVuln (\_ b -> return b) -- noop update functions
(\_ t -> return t)
(\_ e -> return e)
main :: IO ()
main = defaultMainWith
(defaultConfig {reportFile = Just "noop-criterion.html"})
[ bgroup "noop"
[ bench "noopProcessData 20" $ nfIO (noopProcessData "benchmark/sample.csv")
, bench "noopProcessData 200" $ nfIO (noopProcessData "benchmark/sample200.csv")
]
]
| srenatus/avus | benchmark/AvusBenchmark.hs | mit | 576 | 0 | 12 | 125 | 186 | 95 | 91 | 14 | 1 |
{-# LANGUAGE OverloadedStrings #-}
{-# LANGUAGE TemplateHaskell #-}
module Main where
import Control.Applicative ((<|>))
import Text.Trifecta
import Data.ByteString (ByteString)
import qualified Data.ByteString.Char8 as BC
import Data.FileEmbed (embedFile)
input :: ByteString
input = $(embedFile "input.txt")
data Input
= Plain ByteString
| Compressed { _reps :: Int
, _string :: ByteString}
aXb :: Parser (Int, Int)
aXb =
parens $
do a <- natural
_ <- char 'x'
b <- natural
return (fromIntegral a, fromIntegral b)
compressed1 :: Parser Int
compressed1 = do
(c,r) <- aXb
s <- count c anyChar
return $ r * length s
plain :: Parser Int
plain = length <$> some (notChar '(')
parse :: Parser Int -> Parser Int -> ByteString -> [Int]
parse f g b =
case parseByteString (many $ f <|> g) mempty b of
Success is -> is
Failure ei -> fail . show $ ei
decompress1 :: ByteString -> Int
decompress1 = sum . parse compressed1 plain
test1in :: [ByteString]
test1in =
[ "ADVENT"
, "A(1x5)BC"
, "(3x3)XYZ"
, "A(2x2)BCD(2x2)EFG"
, "(6x1)(1x3)A"
, "X(8x2)(3x3)ABCY"]
test1out :: [ByteString]
test1out =
[ "ADVENT"
, "ABBBBBC"
, "XYZXYZXYZ"
, "ABCBCDEFEFG"
, "(1x3)A"
, "X(3x3)ABC(3x3)ABCY"]
test1 :: Bool
test1 = and $ zipWith (==) (map decompress1 test1in) (map BC.length test1out)
part1 :: Int
part1 = decompress1 input
compressed2 :: Parser Int
compressed2 = do
(c,r) <- aXb
s <- count c anyChar
return $ r * (sum . parse compressed2 plain $ BC.pack s)
decompress2 :: ByteString -> Int
decompress2 = sum . parse compressed2 plain
test2in :: [ByteString]
test2in =
[ "(3x3)XYZ"
, "X(8x2)(3x3)ABCY"
, "(27x12)(20x12)(13x14)(7x10)(1x12)A"
, "(25x3)(3x3)ABC(2x3)XY(5x2)PQRSTX(18x9)(3x2)TWO(5x7)SEVEN"]
test2out :: [Int]
test2out =
[BC.length "XYZXYZXYZ", BC.length "XABCABCABCABCABCABCY", 241920, 445]
test2 :: Bool
test2 = and $ zipWith (==) (map decompress2 test2in) test2out
part2 :: Int
part2 = decompress2 input
main :: IO ()
main = do
print test1
print part1
print test2
print part2
| genos/online_problems | advent_of_code_2016/day9/src/Main.hs | mit | 2,172 | 0 | 11 | 512 | 699 | 374 | 325 | 81 | 2 |
{-# LANGUAGE LambdaCase, NamedFieldPuns, OverloadedStrings #-}
module Main (main) where
import Devil.Config
import Devil.Daemons
import Options.Applicative
import qualified Data.Text as T
import qualified Devil.Log as Log
data Params = Params {
configFile :: String
} deriving (Show,Eq)
params' :: Parser Params
params'
= Params
<$> strOption
( long "config"
<> short 'c'
<> metavar "CONFIG"
<> help "Config file")
params :: ParserInfo Params
params = info (helper <*> params')
( progDesc "Small (and silly!) `daemon` manager"
)
main :: IO ()
main = do
Log.logThread
go =<< execParser params where
go (Params{configFile}) =
loadConfig configFile >>= \case
Left err -> do
Log.error_d "CONFIG" (T.pack $ show err)
Right cfg ->
runDaemons cfg
| EXio4/devil | src/Main.hs | mit | 961 | 0 | 16 | 333 | 246 | 128 | 118 | 31 | 2 |
{-# LANGUAGE QuasiQuotes #-}
module Minesweeper where
import Prelude hiding (map, zipWith)
import Data.Maybe (fromJust)
import qualified Data.Vector as V
import Data.Array.Repa as R hiding ((++))
import Data.Array.Repa.Repr.Vector
import Data.Array.Repa.Stencil
import Data.Array.Repa.Stencil.Dim2
import Data.Array.Repa.Algorithms.Randomish
import System.Random
-- the danger rating of a minesweeper square
-- I think that the Danger type is isomorphic to Maybe Integer
-- if it helps me to have a monad/functor/applicative instance
-- for danger, I can just use Maybe's implementation and it should
-- work.
data Danger = Mine | Danger Int deriving Eq
-- a square keeps track of if it is revealed, as well as the danger rating
-- of the location it is in
data Square = Square { getDanger :: Danger
, isRevealed :: Bool
} deriving Eq
emptySquare = Square (Danger 0) False
-- For minesweeper, there are a couple of things that we need to keep track of
-- First, we need the locations of the mines
-- We also need which squares are revealed at any given time
-- the numbers are calculatable, but I think that we should calculate them
-- once at the beginning and then store them
-- the field is row-major
-- note: with the new repa representation, the width and height of the array
-- are encoded in the array as the shape
data Field = Field { getField :: Array V DIM2 Square }
instance Show Danger where
show Mine = "*"
-- any danger level is guarenteed to be 1 digit, because the most neighbors
-- any square can have is 8
--show (Danger 0) = " "
show (Danger d) = show d
instance Show Square where
show (Square d True) = show d
show (Square _ False) = " "
--debug show
--show (Square d _) = show d
instance Show Field where
show (Field f) = let (Z :. x :. y) = extent f in
V.foldl1' (++)
$ V.imap (\i a -> if i /= 0 then (if mod i y == 0 then '\n' else ' ') : show a else show a)
$ toVector f
-- take n unique values from a list. the third argument is the seen list
takeUnique :: (Eq n, Eq a, Num n) => n -> [a] -> [a]
takeUnique = takeUnique' []
takeUnique' :: (Eq n, Eq a, Num n) => [a] -> n -> [a] -> [a]
takeUnique' seen 0 _ = seen
takeUnique' seen n (x:xs) | x `notElem` seen = takeUnique' (x:seen) (n-1) xs
| otherwise = takeUnique' seen n xs
-- generate a new field using one rng for x and one rng for y
generateField :: (RandomGen g)
=> g -- random number generator
-> Int -- height of minefield
-> Int -- width of minefield
-> Int -- Number of mines
-> Field
generateField g x y n = updateDangers $ Field f
where rs = takeUnique n $ randomRs (0, (x * y) - 1) g
fv = V.replicate (x * y) emptySquare V.// zip rs (repeat $ Square Mine False)
f = fromVector (Z :. x :. y) fv
updateDangers :: Field -> Field
updateDangers (Field f) = Field
. computeVectorS -- convert from delayed array back to vector array
. zipWith updateDanger f -- update the field with the new dangers we found
. mapStencil2 (BoundConst 0) stencil -- perform the convolution
. map (\s->if isMine s then 1 else 0) -- turn mines into 1s and others into 0s
$ f
where stencil = [stencil2|1 1 1
1 0 1
1 1 1|]
updateDanger :: Square -> Int -> Square
updateDanger (Square (Danger _) r) d = Square (Danger d) r
updateDanger s _ = s
isMine :: Square -> Bool
isMine (Square Mine _) = True
isMine _ = False
isRevealedMine :: Square -> Bool
isRevealedMine (Square Mine True) = True
isRevealedMine _ = False
isMineOrRevealed :: Square -> Bool
isMineOrRevealed (Square Mine False) = True
isMineOrRevealed (Square (Danger _) True) = True
isMineOrRevealed _ = False
-- reveal spot
-- coordinates are measured from top left of grid
-- the basic reveal is to just set the reveal mask for that spot to be True,
-- and return the Left Field if that spot is a mine, and Right Field if it's
-- not
-- past that, you could reasonably space-fill the revealed area if there are
-- no mines in the surrounding areas, much like most implementations for
-- playing but I think we can leave that task on the consumer of the game,
-- and just have this a "recorder of events" of sorts
reveal :: Int -- x value of spot to reveal
-> Int -- y value of spot to reveal
-> Field -- field to reveal spot on
-> Field -- no computation is done to imply whether or not a mine was hit
reveal x y (Field f) = Field (fromJust (computeP (R.traverse f id (update x y))))
where update x y square s@(Z :. sx :. sy) | x == sx && y == sy = (square s){isRevealed = True}
update x y square s = square s
hitMine :: Field -> Bool
hitMine = fromJust . foldAllP (||) False . R.map isRevealedMine . getField
solved :: Field -> Bool
solved = fromJust . foldAllP (&&) True . R.map isMineOrRevealed . getField
| sdemos/minesweeper | src/Minesweeper.hs | mit | 5,242 | 0 | 18 | 1,505 | 1,229 | 673 | 556 | 76 | 2 |
sumtorial :: Integer -> Integer
sumtorial 0 = 0
sumtorial n = n + sumtorial (n - 1)
| martindavid/code-sandbox | haskell/exercises/exercise1.hs | mit | 84 | 0 | 8 | 18 | 40 | 20 | 20 | 3 | 1 |
module Main where
import Test.Framework (defaultMain)
import Language.Swift.Tests (tests)
main :: IO ()
main = defaultMain
[ tests
]
| CodaFi/language-swift | tests/Tests.hs | mit | 144 | 0 | 6 | 30 | 46 | 27 | 19 | 6 | 1 |
module Reader where
import Control.Applicative ((<$>), (<*>))
import Text.ParserCombinators.Parsec
import Text.ParserCombinators.Parsec.Language
import Text.ParserCombinators.Parsec.Token
import LispData
import Numbers
-- | parses lisp code from a string and returns either the code an error message
reader :: String -> Either String LispVal
reader str = case parse parser "Lisp" str of
Left err -> Left (show err)
Right res -> Right res
where parser = const . makeForms <$> many lispParser <*> eof
makeForms [form] = form
makeForms forms = List (Symbol "begin" : forms)
-- | parses one lisp token
lexer :: TokenParser ()
lexer = makeTokenParser LanguageDef
{ commentStart = "#|"
, commentEnd = "|#"
, commentLine = ";"
, nestedComments = True
, identStart = symChar
, identLetter = symChar
, opStart = oneOf ""
, opLetter = oneOf ""
, reservedNames = []
, reservedOpNames = []
, caseSensitive = False
}
where symChar = alphaNum <|> oneOf "?+*~#-_.:=&%$!^<>|/"
-- | parses lisp
lispParser :: Parser LispVal
lispParser = do
whiteSpace lexer
val <- try numberParser <|>
try boolParser <|>
symbolParser <|>
stringParser <|>
listParser <|>
quoteParser <|>
quasiQuoteParser <|>
unquoteParser
whiteSpace lexer
return val
symbolParser :: Parser LispVal
symbolParser = do
name <- identifier lexer
return $ if name == "nil"
then Nil
else Symbol name
-- TODO: nice literals for rationals and complex numbers
numberParser :: Parser LispVal
numberParser = do
sign <- optionMaybe (oneOf "-+")
let mySign = case sign of
Just '-' -> -1
_ -> 1
number <- naturalOrFloat lexer
return $ Number $ case number of
Left i -> LispInt (mySign * i)
Right f -> LispFloat (fromIntegral mySign * f)
stringParser :: Parser LispVal
stringParser = LispString <$> stringLiteral lexer
listParser :: Parser LispVal
listParser = (\_ content _ -> List content) <$> char '(' <*> many lispParser <*> char ')'
quoteParser :: Parser LispVal
quoteParser = (\_ val -> List [Symbol "quote", val]) <$> char '\'' <*> lispParser
quasiQuoteParser :: Parser LispVal
quasiQuoteParser = (\_ val -> List [Symbol "quasiquote", val]) <$> char '`' <*> lispParser
unquoteParser :: Parser LispVal
unquoteParser = (\_ val -> List [Symbol "unquote", val]) <$> char ',' <*> lispParser
boolParser :: Parser LispVal
boolParser = char '#' >> ((char 't' >> return (Boolean True)) <|> (char 'f' >> return (Boolean False)))
| orion-42/my-lisp | Reader.hs | mit | 2,630 | 0 | 15 | 648 | 792 | 408 | 384 | 69 | 3 |
{-# LANGUAGE DeriveDataTypeable #-}
module Base.CLI (ProgramOptions(..), Action(..), usage, newCommands, standard, module System.Console.CmdArgs) where
import System.Console.CmdArgs
import Base.Common
data Action = Format | ListCommands
deriving (Show, Typeable, Data)
instance Default Action where
def = Format
data ProgramOptions = ProgramOptions
{ agda_mode :: Bool
, action :: Action
, input :: [FilePath]
, output :: FilePath
} deriving (Show, Data, Typeable)
usage :: String
usage = unlines
[ programName ++" "++ programVersion ++" - A lhs2TeX Syntax Colouring preprocessor"
, "Consult the README file for extra information or visit:\n"
, " https://github.com/spockz/lhs2texhl"
, " and "
, " http://alessandrovermeulen.me/projects/lhs2texhl\n"
, "Copyright 2010, Alessandro Vermeulen <[email protected]>" ]
newCommands :: String
newCommands = unlines [
"\\newcommand{\\lhsCHfunction}[1]{\\color{infixoperator}{{#1}}}",
"\\newcommand{\\lhsCHinfixoperator}[1]{\\color{infixoperator}{{#1}}}",
"\\newcommand{\\lhsCHprelude}[1]{\\color{prelude}{{#1}}}",
"\\newcommand{\\lhsCHkeyword}[1]{\\color{keyword}{{#1}}}",
"\\newcommand{\\lhsCHconstructor}[1]{\\color{constructor}{{#1}}}",
"\\newcommand{\\lhsCHtype}[1]{\\color{datatype}{{#1}}}",
"\\newcommand{\\lhsCHsyntax}[1]{\\color{syntax}{{#1}}}",
"\\newcommand{\\lhsCHclass}[1]{\\color{class}{{#1}}}",
"\\newcommand{\\lhsCHconstant}[1]{\\color{constant}{{#1}}}"
]
-- | Standard command line options.
--
standard = cmdArgsMode $ ProgramOptions
{
agda_mode = def &= help "Run in agda-mode!"
, action = (def &= help "What should the program do? Format|ListCommands.") &= typ "Action"
, output = (def &= help "Output file") &= typFile
, input = (def &= args )
} &= summary usage
| spockz/lhs2texhl | src/Base/CLI.hs | mit | 1,973 | 0 | 12 | 421 | 311 | 185 | 126 | 39 | 1 |
{-# LANGUAGE CPP #-}
module GHCJS.DOM.SVGFESpotLightElement (
#if (defined(ghcjs_HOST_OS) && defined(USE_JAVASCRIPTFFI)) || !defined(USE_WEBKIT)
module GHCJS.DOM.JSFFI.Generated.SVGFESpotLightElement
#else
#endif
) where
#if (defined(ghcjs_HOST_OS) && defined(USE_JAVASCRIPTFFI)) || !defined(USE_WEBKIT)
import GHCJS.DOM.JSFFI.Generated.SVGFESpotLightElement
#else
#endif
| plow-technologies/ghcjs-dom | src/GHCJS/DOM/SVGFESpotLightElement.hs | mit | 376 | 0 | 5 | 33 | 33 | 26 | 7 | 4 | 0 |
module GraphDB.Util.Prelude.TH
(
module Exports,
purify,
tryToReify,
isInstance',
isProperInstance',
)
where
import GraphDB.Util.Prelude hiding (Fixity)
import Language.Haskell.TH as Exports
import Language.Haskell.TH.Syntax as Exports
import THInstanceReification as Exports
purify :: Q a -> a
purify = unsafePerformIO . runQ
tryToReify :: Name -> Q (Maybe Info)
tryToReify n = recover (return Nothing) (fmap Just $ reify n)
isInstance' :: Name -> [Type] -> Q Bool
isInstance' name types = recover (return False) (isInstance name types)
isProperInstance' :: Name -> [Type] -> Q Bool
isProperInstance' name types = recover (return False) (isProperInstance name types)
| nikita-volkov/graph-db | library/GraphDB/Util/Prelude/TH.hs | mit | 686 | 0 | 8 | 110 | 228 | 127 | 101 | 19 | 1 |
{-# LANGUAGE FlexibleContexts #-}
{-# LANGUAGE ScopedTypeVariables #-}
module Server.Main where
import System.Directory
import qualified Text.Blaze.Html.Renderer.Text as H
import Text.Pandoc
import System.FilePath
import qualified Data.ByteString as B
import qualified Data.ByteString.Lazy as BL
import qualified Data.Text as T
import qualified Data.Text.Lazy as LT
import qualified Data.Text.Lazy.Encoding as LT
import qualified Data.Set as S
import Text.Pandoc.Walk
import Control.Monad.Reader
import Config
import Utils
import API
import Servant
import Control.Monad.Error.Class
import Text.Pandoc.CrossRef
import Data.Monoid ((<>))
import Text.Pandoc.Builder
import Data.Generics
import Crypto.Hash (hash, Digest, SHA1)
import Data.Char (isAlphaNum)
import Control.Exception
import System.IO.Error
mainServer :: ServerT MainAPI ConfigHandler
mainServer (_ :: User)
= listProjects
:<|> createProject
:<|> deleteProject
:<|> render
:<|> update
:<|> appendChunk
:<|> getSource
:<|> fileList
:<|> deleteFile
:<|> uploadFile
:<|> renderDocx
mdOpts :: WriterOptions
mdOpts = def {
writerSetextHeaders = False
, writerExtensions = S.delete Ext_simple_tables pandocExtensions
}
htmlOpts :: WriterOptions
htmlOpts = def{
writerHtml5 = True
, writerHTMLMathMethod = MathJax ""
}
handlePandocError :: (MonadError ServantErr m) => Either PandocError Pandoc -> m Pandoc
handlePandocError (Left err) = throwError err500{ errBody = LT.encodeUtf8 $ LT.pack $ show err }
handlePandocError (Right res) = return res
getBody :: (MonadIO m, MonadError ServantErr m, MonadReader Config m) => FilePath -> m Pandoc
getBody name = do
validateName name
dataDirectory <- asks configDataDir
everywhere (mkT splitMath) <$> (handlePandocError . readMarkdown def
=<< liftIO (readFile (dataDirectory </> name </> "index.md")))
splitMath :: [Block] -> [Block]
splitMath (Para ils:xs)
| length ils > 1 = map Para (split ils) ++ xs
where
split ys =
let bef = takeWhile (not . isMath) ys
rest = drop (length bef) ys
m = takeWhile (not . isSpace) rest
af = drop (length m) rest
in filter (not . null) [bef, m, af]
isMath (Math DisplayMath _) = True
isMath (Span _ [Math DisplayMath _]) = True
isMath _ = False
isSpace Space = True
isSpace SoftBreak = True
isSpace _ = False
splitMath xs = xs
createProject :: FilePath -> T.Text -> ConfigHandler ()
createProject name content = do
projects <- liftIO . listDirectory =<< asks configDataDir
when (name `elem` projects) $ throwError err409
unless (all isAlphaNum name) $ throwError err400
dataDir <- asks configDataDir
liftIO $ do
createDirectory (dataDir </> name)
BL.writeFile (dataDir </> name </> "index.md") $ LT.encodeUtf8 $ LT.fromStrict content
deleteProject :: FilePath -> ConfigHandler ()
deleteProject name = do
validateName name
dataDir <- asks configDataDir
liftIO $ removeDirectoryRecursive (dataDir </> name)
render :: FilePath -> ConfigHandler [Chunk]
render name = do
Pandoc meta body <- getBody name
uri <- asks configDataUri
let modImgs (Image a t (src, tit)) = Image a t (uri </> name </> src, tit)
modImgs x = x
let body' = runCrossRef (crossRefSettings <> meta) Nothing crossRefBlocks . wrapDiv $ walk modImgs body
return $ zipWith3 (mkChunk meta) body body' [0..]
where
wrapDiv = map (Div nullAttr . return)
mkChunk meta blold bl idx = Chunk {
chunkHtml = H.renderHtml $ writeHtml htmlOpts $ Pandoc meta [bl]
, chunkSrc = T.pack $ writeMarkdown mdOpts $ Pandoc meta [blold]
, chunkNum = idx
}
renderDocx :: FilePath -> ConfigHandler FileData
renderDocx name = do
validateName name
dataDirectory <- asks configDataDir
newcmdFile <- asks configNewCmdFile
newcommands <-
either (const "") id <$>
maybe (return $ Left ())
(liftIO . tryJust (guard . isDoesNotExistError) . readFile)
newcmdFile
Pandoc meta body <- handlePandocError . readMarkdown def . (newcommands <>)
=<< liftIO (readFile (dataDirectory </> name </> "index.md"))
let modImgs (Image a t (src, tit)) = Image a t (dataDirectory </> name </> src, tit)
modImgs x = x
let body' = runCrossRef (crossRefSettings <> meta) Nothing crossRefBlocks $ walk modImgs body
FileData <$> liftIO (writeDocx def (Pandoc meta body'))
crossRefSettings :: Meta
crossRefSettings =
chapters True
<> numberSections True
<> sectionsDepth "3"
<> chaptersDepth "1"
<> figureTitle (str "Рисунок")
<> tableTitle (str "Таблица")
<> listingTitle (str "Листинг")
<> figPrefix [str "рис."]
<> eqnPrefixTemplate (str "(" <> var "i" <> str ")")
<> tblPrefix [str "табл."]
<> lstPrefix [str "лист."]
<> secPrefix [str "разд."]
<> lofTitle (header 1 $ text "Список рисунков")
<> lotTitle (header 1 $ text "Список таблиц")
<> lolTitle (header 1 $ text "Список листингов")
-- <> autoEqnLabels True
<> subfigGrid True
<> linkReferences True
where var = displayMath
update :: FilePath -> Int -> T.Text -> ConfigHandler ()
update name chunk mdbody = do
Pandoc meta body <- getBody name
dataDirectory <- asks configDataDir
validateChunk chunk (length body)
Pandoc _ newChunkBody <- handlePandocError $ readMarkdown def $ T.unpack mdbody
let (b1, _:b2) = splitAt chunk body
body' = b1 ++ newChunkBody ++ b2
liftIO
$ writeFile (dataDirectory </> name </> "index.md")
$ writeMarkdown mdOpts $ Pandoc meta body'
appendChunk :: FilePath -> T.Text -> ConfigHandler ()
appendChunk name mdbody = do
Pandoc meta body <- getBody name
dataDirectory <- asks configDataDir
Pandoc _ newChunkBody <- handlePandocError $ readMarkdown def $ T.unpack mdbody
let body' = body ++ newChunkBody
liftIO
$ writeFile (dataDirectory </> name </> "index.md")
$ writeMarkdown mdOpts $ Pandoc meta body'
getSource :: FilePath -> Int -> ConfigHandler T.Text
getSource name chunk = do
Pandoc meta body <- getBody name
validateChunk chunk (length body)
let c = [body !! chunk]
return $ T.pack $ writeMarkdown mdOpts (Pandoc meta c)
listProjects :: ConfigHandler [FilePath]
listProjects = do
dataDirectory <- asks configDataDir
liftIO $ listDirectory dataDirectory
uploadFile :: FilePath -> FileData -> ConfigHandler T.Text
uploadFile name (FileData content) = do
validateName name
dataDirectory <- asks configDataDir
let content' = BL.toStrict content
filename = show (hash content' :: Digest SHA1)
liftIO $ B.writeFile (dataDirectory </> name </> filename) content'
return $ T.pack filename
fileList :: FilePath -> ConfigHandler [FileInfo]
fileList name = do
validateName name
dataDirectory <- asks configDataDir
uriBase <- asks configDataUri
map (mkFI uriBase) . filter (/= "index.md") <$> liftIO (listDirectory (dataDirectory </> name))
where
mkFI uriBase fn = FileInfo {
fileName = T.pack fn
, fileURI = T.pack $ uriBase </> name </> fn
}
deleteFile :: FilePath -> FilePath -> ConfigHandler ()
deleteFile proj fn = do
validateFile proj fn
dataDirectory <- asks configDataDir
liftIO $ removeFile (dataDirectory </> proj </> fn)
| lierdakil/markco | server/src/Server/Main.hs | mit | 7,255 | 0 | 21 | 1,428 | 2,497 | 1,237 | 1,260 | 189 | 5 |
{-# LANGUAGE OverloadedStrings #-}
module HailsRock.Views where
import Prelude hiding (div, span, head, id)
import Data.Maybe (isJust, fromJust)
import qualified Data.ByteString.Lazy.Char8 as L8
import qualified Data.Text as T
import Text.Blaze.Html5 hiding (Tag, map)
import Text.Blaze.Html5.Attributes hiding ( label, form, span
, title, style )
import qualified Text.Blaze.Html5.Attributes as A
import Text.Blaze.Html.Renderer.Utf8
import Control.Monad (forM_, when)
import Hails.Web hiding (body)
import Hails.HttpServer.Types
import HailsRock.MP
respondHtml :: Html -> Response
respondHtml content = okHtml $ renderHtml $ docTypeHtml $ do
head $ do
title "HailsRock"
meta ! charset "utf-8"
link ! rel "stylesheet"
! type_ "text/css" ! href "/static/css/bootstrap.css"
script ! src "/static/js/jquery-1.10.1.js" $ ""
script ! src "/static/js/bootstrap.js" $ ""
script ! src "/static/js/application.js" $ ""
body $ do
div ! class_ "container-fluid" $ content
welcome :: Maybe UserName -> Html
welcome Nothing = do
h1 $ "Welcome to HailsRock!"
a ! class_ "btn btn-large btn-info"
! href "/login"
$ "Login to play"
welcome (Just usr) = do
h1 $ toHtml $ "Welcome to HailsRock, " ++ T.unpack usr ++ "!"
a ! class_ "btn btn-large btn-primary"
! href "/game/new"
$ "Create a new game"
" "
a ! class_ "btn btn-large"
! href "/game"
$ "Join a game"
newGame :: UserName -> Html
newGame usr = do
h1 $ "Create a new game"
div $ do
form ! action "/game/create" ! method "POST" ! id "newGame"$ do
div $ do
input ! type_ "hidden" ! name "creator"
! value (toValue usr)
div $ do
label ! for "opponent" $ "Opponent (optional):"
input ! type_ "text"
! name "opponent" ! id "opponent"
! placeholder "rick-james"
div ! class_ "btn-group" $ do
input ! type_ "submit" ! class_ "btn" ! value "Create"
listGames :: UserName -> [Game] -> Html
listGames usr gs' = do
-- Get all the games for which the current user is not the creator;
let gs = filter ((/= usr) . creator) gs'
--
h1 $ "Available games"
div $ if null gs
then p $ "Sorry, no games ... :-("
else table ! class_ "table table-hover table-condensed" $ do
thead $ tr $ do
th $ "#"
th $ "Creator"
th $ "Private"
tbody $ do
forM_ (zip [1..] gs) $ \(nr,game) -> do
let tagUrl = "/game/" ++ show (fromJust $ gameId game)
tr ! onclick (toValue $ "location.href=" ++ show tagUrl )$ do
td $ toHtml (nr :: Int)
td $ toHtml $ creator game
td $ when (isJust $ opponent game) $ "1-vs-1"
playGame :: UserName -> Game -> Bool -> Html
playGame usr game True = do
h1 $ "You already played!"
playGame usr game False = do
h1 $ "Make your move..."
div $ do
let gid = show . fromJust . gameId $ game
form ! action (toValue $ "/game/"++gid++"/play")
! method "POST" ! id "newGame"$ do
input ! type_ "hidden" ! name "game"
! value (toValue gid)
input ! type_ "hidden" ! name "player"
! value (toValue usr)
input ! name "move"
! type_ "submit"
! class_ "btn btn-large btn-info"
! value (toValue $ show Rock)
" "
input ! name "move"
! type_ "submit"
! class_ "btn btn-large btn-primary"
! value (toValue $ show Paper)
" "
input ! name "move"
! type_ "submit"
! class_ "btn btn-large btn-inverse"
! value (toValue $ show Scissors)
showStats :: [(UserName, Outcome)] -> Html
showStats stats = do
h1 $ "Your move status"
div $ if null stats
then p $ "Sorry, nobody has played your move... :-("
else table ! class_ "table table-hover table-condensed" $ do
thead $ tr $ do
th $ "#"
th $ "Player"
th $ "Status"
tbody $ do
forM_ (zip [1..] stats) $ \(nr,(p,result)) -> do
tr $ do
td $ toHtml (nr :: Int)
td $ toHtml $ T.unpack p
td $ toHtml $ show result
| scslab/hails | examples/hails-rock/HailsRock/Views.hs | mit | 4,365 | 0 | 27 | 1,434 | 1,388 | 673 | 715 | 118 | 2 |
main :: IO()
main = putStrLn "Hello world!"
add a b = a + b
x = 10
myDrop n xs = if n <= 0 || null xs
then xs
else myDrop (n-1) (tail xs)
data BookInfo = Book Int String [String]
deriving (Show)
data MagzineInfo = Magzine Int String [String]
deriving (Show)
myInfo = Book 9780135072455 "Algebra of Programming"
["Richard Bird", "Oege de Moor"]
type CustomerID = Int
type ReviewBody = String
data BookReview = BookReview BookInfo CustomerID String
type BookRecord = (BookInfo, BookReview)
data Cartesian2D = Cartesian2D Double Double deriving (Eq, Show)
data Polar2D = Polar2D Double Double deriving (Eq, Show)
data Roygbiv = Red
| Orange
| Yellow
| Green
| Blue
| Infigo
| Violet deriving (Eq, Show)
type Vector = (Double, Double)
data Shape = Circle Vector Double
| Ploy [Vector] deriving (Show)
myNot True = False
myNot False = True
sumList (x:xs) = x + sumList xs
sumList [] = 0
third (a, b, c) = c
complicated (True, a, x:xs, 5) = (a, xs)
bookID (Book id title authors) = id
bookTitle (Book id title authors) = title
bookAuthors (Book id title authors) = authors
nicerID (Book id _ _ ) = id
nicerTitle (Book _ title _ ) = title
nicerAuthors (Book _ _ authors ) = authors
isTwo :: Int -> Bool
isTwo n = if n == 2 then True else False
month :: Int -> Int
month = undefined
days :: (Int, Int) -> Int
days (m, d) = month m + d
let xxx = concat $ map
| dalonng/hellos | haskell.hello/hello.hs | gpl-2.0 | 1,572 | 1 | 8 | 483 | 599 | 327 | 272 | -1 | -1 |
{-
Copyright (C) 2014 Jesse Rosenthal <[email protected]>
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 2 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program; if not, write to the Free Software
Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
-}
{- |
Module : Text.Pandoc.Readers.DocX.Lists
Copyright : Copyright (C) 2014 Jesse Rosenthal
License : GNU GPL, version 2 or above
Maintainer : Jesse Rosenthal <[email protected]>
Stability : alpha
Portability : portable
Functions for converting flat DocX paragraphs into nested lists.
-}
module Text.Pandoc.Readers.DocX.Lists ( blocksToBullets
, blocksToDefinitions) where
import Text.Pandoc.JSON
import Text.Pandoc.Shared (trim)
import Control.Monad
import Data.List
import Data.Maybe
isListItem :: Block -> Bool
isListItem (Div (_, classes, _) _) | "list-item" `elem` classes = True
isListItem _ = False
getLevel :: Block -> Maybe Integer
getLevel (Div (_, _, kvs) _) = liftM read $ lookup "level" kvs
getLevel _ = Nothing
getLevelN :: Block -> Integer
getLevelN b = case getLevel b of
Just n -> n
Nothing -> -1
getNumId :: Block -> Maybe Integer
getNumId (Div (_, _, kvs) _) = liftM read $ lookup "num-id" kvs
getNumId _ = Nothing
getNumIdN :: Block -> Integer
getNumIdN b = case getNumId b of
Just n -> n
Nothing -> -1
getText :: Block -> Maybe String
getText (Div (_, _, kvs) _) = lookup "text" kvs
getText _ = Nothing
data ListType = Itemized | Enumerated ListAttributes
listStyleMap :: [(String, ListNumberStyle)]
listStyleMap = [("upperLetter", UpperAlpha),
("lowerLetter", LowerAlpha),
("upperRoman", UpperRoman),
("lowerRoman", LowerRoman),
("decimal", Decimal)]
listDelimMap :: [(String, ListNumberDelim)]
listDelimMap = [("%1)", OneParen),
("(%1)", TwoParens),
("%1.", Period)]
getListType :: Block -> Maybe ListType
getListType b@(Div (_, _, kvs) _) | isListItem b =
let
start = lookup "start" kvs
frmt = lookup "format" kvs
txt = lookup "text" kvs
in
case frmt of
Just "bullet" -> Just Itemized
Just f ->
case txt of
Just t -> Just $ Enumerated (
read (fromMaybe "1" start) :: Int,
fromMaybe DefaultStyle (lookup f listStyleMap),
fromMaybe DefaultDelim (lookup t listDelimMap))
Nothing -> Nothing
_ -> Nothing
getListType _ = Nothing
listParagraphDivs :: [String]
listParagraphDivs = ["ListParagraph"]
-- This is a first stab at going through and attaching meaning to list
-- paragraphs, without an item marker, following a list item. We
-- assume that these are paragraphs in the same item.
handleListParagraphs :: [Block] -> [Block]
handleListParagraphs [] = []
handleListParagraphs (
(Div attr1@(_, classes1, _) blks1) :
(Div (ident2, classes2, kvs2) blks2) :
blks
) | "list-item" `elem` classes1 &&
not ("list-item" `elem` classes2) &&
(not . null) (listParagraphDivs `intersect` classes2) =
-- We don't want to keep this indent.
let newDiv2 =
(Div (ident2, classes2, filter (\kv -> fst kv /= "indent") kvs2) blks2)
in
handleListParagraphs ((Div attr1 (blks1 ++ [newDiv2])) : blks)
handleListParagraphs (blk:blks) = blk : (handleListParagraphs blks)
separateBlocks' :: Block -> [[Block]] -> [[Block]]
separateBlocks' blk ([] : []) = [[blk]]
separateBlocks' b@(BulletList _) acc = (init acc) ++ [(last acc) ++ [b]]
separateBlocks' b@(OrderedList _ _) acc = (init acc) ++ [(last acc) ++ [b]]
-- The following is for the invisible bullet lists. This is how
-- pandoc-generated ooxml does multiparagraph item lists.
separateBlocks' b acc | liftM trim (getText b) == Just "" =
(init acc) ++ [(last acc) ++ [b]]
separateBlocks' b acc = acc ++ [[b]]
separateBlocks :: [Block] -> [[Block]]
separateBlocks blks = foldr separateBlocks' [[]] (reverse blks)
flatToBullets' :: Integer -> [Block] -> [Block]
flatToBullets' _ [] = []
flatToBullets' num xs@(b : elems)
| getLevelN b == num = b : (flatToBullets' num elems)
| otherwise =
let bNumId = getNumIdN b
bLevel = getLevelN b
(children, remaining) =
span
(\b' ->
((getLevelN b') > bLevel ||
((getLevelN b') == bLevel && (getNumIdN b') == bNumId)))
xs
in
case getListType b of
Just (Enumerated attr) ->
(OrderedList attr (separateBlocks $ flatToBullets' bLevel children)) :
(flatToBullets' num remaining)
_ ->
(BulletList (separateBlocks $ flatToBullets' bLevel children)) :
(flatToBullets' num remaining)
flatToBullets :: [Block] -> [Block]
flatToBullets elems = flatToBullets' (-1) elems
blocksToBullets :: [Block] -> [Block]
blocksToBullets blks =
-- bottomUp removeListItemDivs $
flatToBullets $ (handleListParagraphs blks)
plainParaInlines :: Block -> [Inline]
plainParaInlines (Plain ils) = ils
plainParaInlines (Para ils) = ils
plainParaInlines _ = []
blocksToDefinitions' :: [([Inline], [[Block]])] -> [Block] -> [Block] -> [Block]
blocksToDefinitions' [] acc [] = reverse acc
blocksToDefinitions' defAcc acc [] =
reverse $ (DefinitionList (reverse defAcc)) : acc
blocksToDefinitions' defAcc acc
((Div (_, classes1, _) blks1) : (Div (ident2, classes2, kvs2) blks2) : blks)
| "DefinitionTerm" `elem` classes1 && "Definition" `elem` classes2 =
let remainingAttr2 = (ident2, delete "Definition" classes2, kvs2)
pair = case remainingAttr2 == ("", [], []) of
True -> (concatMap plainParaInlines blks1, [blks2])
False -> (concatMap plainParaInlines blks1, [[Div remainingAttr2 blks2]])
in
blocksToDefinitions' (pair : defAcc) acc blks
blocksToDefinitions' defAcc acc
((Div (ident2, classes2, kvs2) blks2) : blks)
| (not . null) defAcc && "Definition" `elem` classes2 =
let remainingAttr2 = (ident2, delete "Definition" classes2, kvs2)
defItems2 = case remainingAttr2 == ("", [], []) of
True -> blks2
False -> [Div remainingAttr2 blks2]
((defTerm, defItems):defs) = defAcc
defAcc' = case null defItems of
True -> (defTerm, [defItems2]) : defs
False -> (defTerm, init defItems ++ [last defItems ++ defItems2]) : defs
in
blocksToDefinitions' defAcc' acc blks
blocksToDefinitions' [] acc (b:blks) =
blocksToDefinitions' [] (b:acc) blks
blocksToDefinitions' defAcc acc (b:blks) =
blocksToDefinitions' [] (b : (DefinitionList (reverse defAcc)) : acc) blks
blocksToDefinitions :: [Block] -> [Block]
blocksToDefinitions = blocksToDefinitions' [] []
| uws-eresearch/docx2pandoc | src/Text/Pandoc/Readers/DocX/Lists.hs | gpl-2.0 | 7,237 | 0 | 20 | 1,684 | 2,211 | 1,186 | 1,025 | 138 | 4 |
{-# LANGUAGE LambdaCase #-}
{-# LANGUAGE OverloadedStrings #-}
{-# LANGUAGE TupleSections #-}
{-# OPTIONS_HADDOCK show-extensions #-}
-- |
-- Module : Yi.Keymap.Vim.Ex.Commands.Quit
-- License : GPL-2
-- Maintainer : [email protected]
-- Stability : experimental
-- Portability : portable
--
-- Implements quit commands.
module Yi.Keymap.Vim.Ex.Commands.Quit (parse) where
import Control.Applicative
import Control.Lens
import Control.Monad
import Data.Foldable (find)
import qualified Data.List.PointedList.Circular as PL
import Data.Monoid
import qualified Data.Text as T
import qualified Text.ParserCombinators.Parsec as P
import Yi.Buffer
import Yi.Core (quitEditor, errorEditor, closeWindow)
import Yi.Editor
import Yi.File
import Yi.Keymap
import Yi.Keymap.Vim.Common
import qualified Yi.Keymap.Vim.Ex.Commands.Common as Common
import Yi.Keymap.Vim.Ex.Types
import Yi.Monad
import Yi.String (showT)
import Yi.Window (bufkey)
parse :: EventString -> Maybe ExCommand
parse = Common.parse $ P.choice
[ do
_ <- (P.try ( P.string "xit") <|> P.string "x")
bangs <- P.many (P.char '!')
return (quit True (not $ null bangs) False)
, do
ws <- P.many (P.char 'w')
void $ P.try ( P.string "quit") <|> P.string "q"
as <- P.many (P.try ( P.string "all") <|> P.string "a")
bangs <- P.many (P.char '!')
return $! quit (not $ null ws) (not $ null bangs) (not $ null as)
]
quit :: Bool -> Bool -> Bool -> ExCommand
quit w f a = Common.impureExCommand {
cmdShow = (if w then "w" else "")
`T.append` "quit"
`T.append` (if a then "all" else "")
`T.append` (if f then "!" else "")
, cmdAction = YiA $ action w f a
}
action :: Bool -> Bool -> Bool -> YiM ()
action False False False = quitWindowE
action False False True = quitAllE
action True False False = viWrite >> closeWindow
action True False True = saveAndQuitAllE
action False True False = closeWindow
action False True True = quitEditor
action True True False = viWrite >> closeWindow
action True True True = saveAndQuitAllE
quitWindowE :: YiM ()
quitWindowE = do
nw <- gets currentBuffer >>= Common.needsSaving
ws <- withEditor $ use currentWindowA >>= windowsOnBufferE . bufkey
if length ws == 1 && nw
then errorEditor "No write since last change (add ! to override)"
else do
winCount <- withEditor $ uses windowsA PL.length
tabCount <- withEditor $ uses tabsA PL.length
if winCount == 1 && tabCount == 1
-- if its the last window, quitting will quit the editor
then quitAllE
else closeWindow
quitAllE :: YiM ()
quitAllE = do
let needsWindow b = (b,) <$> deservesSave b
bs <- readEditor bufferSet >>= mapM needsWindow
-- Vim only shows the first modified buffer in the error.
case find snd bs of
Nothing -> quitEditor
Just (b, _) -> do
bufferName <- withEditor $ withGivenBuffer (bkey b) $ gets file
errorEditor $ "No write since last change for buffer "
<> showT bufferName
<> " (add ! to override)"
saveAndQuitAllE :: YiM ()
saveAndQuitAllE = Common.forAllBuffers fwriteBufferE >> quitEditor | atsukotakahashi/wi | src/library/Yi/Keymap/Vim/Ex/Commands/Quit.hs | gpl-2.0 | 3,438 | 0 | 17 | 953 | 969 | 513 | 456 | 77 | 4 |
-- A QDSL implementation of Carlson, Hudak, and Jones'
-- Geometric Region Servers,
-- by Josef Svenningsson
module Examples.Region where
import QHaskell
type Point = (Float,Float)
type Region = Point -> Bool
type Radius = Float
circle :: Radius -> Region
circle r = \p -> magnitude p <= r
outside :: Region -> Region
outside r = \p -> not (r p)
intersection :: Region -> Region -> Region
intersection r1 r2 = \p -> r1 p && r2 p
union :: Region -> Region -> Region
union r1 r2 = \p -> r1 p || r2 p
magnitude :: Point -> Float
magnitude (x,y) = sqrt (x ** 2 + y ** 2)
makeQDSL "RegionLang" ['circle,'outside,'intersection,'union,'magnitude]
{-
\begin{Code generated automatically by a Template Haskell function}
type Types = [Radius -> Region
,Region -> Region
,Region -> Region -> Region
,Region -> Region -> Region
,Point -> Float]
typeEnv :: TypeEnv Types
typeEnv = 'circle <:> 'outside <:> 'intersection <:> 'union <:>
'magnitude <:> nil
evalEnv :: EvalEnv Types
evalEnv = circle <+> outside <+> intersection <+> union <+>
magnitude <+> nil
translate :: Type a => Qt a -> ErrM (RegionLang a)
translate = tran typeEnv
evaluate :: Type a => RegionLang a -> a
evaluate = eval evalEnv
normalise :: Type a => RegionLang a -> RegionLang a
normalise = norm
type RegionLang a = Dp Types a
qqRegionLang :: Type a => Qt a -> Qt (RegionLang a)
qqRegionLang = tranQ typeEnv
\end{Code generated automatically by a Template Haskell function}
-}
type Result = Region
compile :: Type a => RegionLang a -> a
compile = evaluate
regionLang :: Qt Region -> ErrM Region
regionLang q = do d <- translate q
return (compile (normalise True d))
inRegion :: Qt Region -> Point -> ErrM Bool
inRegion q p = do f <- regionLang q
return (f p)
{-
inRegion :: Qt (Point -> Region -> Bool)
inRegion = [|| \ p r -> r p ||]
-}
test :: ErrM Bool
test = inRegion [|| outside (circle 3) `intersection` circle 6 ||]
(4,4)
| shayan-najd/QHaskell | Examples/Region.hs | gpl-3.0 | 2,038 | 2 | 11 | 500 | 428 | 227 | 201 | -1 | -1 |
module Lamdu.GUI.TagPane
( make
) where
import qualified Control.Lens as Lens
import Data.Binary.Extended (encodeS)
import qualified Data.Char as Char
import Data.Property (Property(..), pVal)
import qualified Data.Property as Property
import qualified Data.Set as Set
import qualified Data.Text as Text
import qualified GUI.Momentu as M
import GUI.Momentu.Align (TextWidget, Aligned(..), WithTextPos(..))
import qualified GUI.Momentu.Align as Align
import qualified GUI.Momentu.Element as Element
import qualified GUI.Momentu.EventMap as E
import GUI.Momentu.Glue ((/-/), (/|/), hbox)
import qualified GUI.Momentu.I18N as MomentuTexts
import GUI.Momentu.ModKey (noMods)
import qualified GUI.Momentu.ModKey as ModKey
import qualified GUI.Momentu.State as GuiState
import GUI.Momentu.Widget (Widget)
import qualified GUI.Momentu.Widget as Widget
import qualified GUI.Momentu.Widgets.DropDownList as DropDownList
import qualified GUI.Momentu.Widgets.FocusDelegator as FocusDelegator
import qualified GUI.Momentu.Widgets.Grid as Grid
import qualified GUI.Momentu.Widgets.Spacer as Spacer
import qualified GUI.Momentu.Widgets.TextEdit as TextEdit
import qualified GUI.Momentu.Widgets.TextEdit.Property as TextEdits
import qualified GUI.Momentu.Widgets.TextView as TextView
import qualified Lamdu.CharClassification as Chars
import qualified Lamdu.Config as Config
import qualified Lamdu.Config.Theme.TextColors as TextColors
import Lamdu.Data.Tag (TextsInLang(..))
import qualified Lamdu.Data.Tag as Tag
import Lamdu.Formatting (Format(..))
import Lamdu.GUI.Styled (addValFrame, label, info, withColor)
import qualified Lamdu.I18N.CodeUI as Texts
import Lamdu.I18N.LangId (LangId(..), _LangId)
import qualified Lamdu.Sugar.Types as Sugar
import Lamdu.Prelude
tagRenameId :: Widget.Id -> Widget.Id
tagRenameId = (`Widget.joinId` ["rename"])
disallowedNameChars :: Set Char
disallowedNameChars = Set.fromList ",[]\\`()"
makeTagNameEdit :: _ => Property f Text -> Widget.Id -> m (TextWidget f)
makeTagNameEdit prop myId =
TextEdits.makeWordEdit
?? pure " "
?? prop
?? tagRenameId myId
<&> Align.tValue . Widget.eventMapMaker . Lens.mapped %~ E.filterChars (`Set.notMember` disallowedNameChars)
makeSymbolNameEdit :: _ => Property f Text -> Widget.Id -> m (TextWidget f)
makeSymbolNameEdit prop myId =
TextEdits.makeWordEdit
<*> (Lens.view (has . Texts.typeOperatorHere) <&> pure)
?? prop
?? tagRenameId myId
<&> Align.tValue . Widget.eventMapMaker . Lens.mapped %~ E.filterChars allowedSymbolChars
where
allowedSymbolChars =
Set.member
?? Set.fromList Chars.operator `Set.difference` disallowedNameChars
makeFocusableTagNameEdit :: _ => Widget.Id -> Property o Text -> m (TextWidget o)
makeFocusableTagNameEdit myId prop =
do
env <- Lens.view id
let fdConfig =
FocusDelegator.Config
{ FocusDelegator.focusChildKeys = env ^. has . Config.jumpToDefinitionKeys
, FocusDelegator.focusChildDoc =
E.toDoc env
[ has . MomentuTexts.edit
, has . Texts.tag
, has . Texts.renameTag
]
, FocusDelegator.focusParentKeys =
[ noMods ModKey.Key'Escape
, noMods ModKey.Key'Enter
]
, FocusDelegator.focusParentDoc =
E.toDoc env
[ has . MomentuTexts.edit
, has . Texts.tag
, has . Texts.stopEditing
]
}
(FocusDelegator.make ?? fdConfig ?? FocusDelegator.FocusEntryParent ?? myId
<&> (Align.tValue %~))
<*> makeTagNameEdit prop myId
makeLanguageTitle :: _ => Widget.Id -> LangId -> m (TextWidget o)
makeLanguageTitle myId lang =
TextView.make
<*> (Lens.view has <&> getLang)
<*> pure (Widget.toAnimId myId <> ["lang-title"])
<&> Align.tValue %~ Widget.fromView
where
getLang :: Map LangId Text -> Text
getLang x =
x ^. Lens.at lang
& fromMaybe (lang ^. _LangId & Lens.ix 0 %~ Char.toUpper)
data TextsRow a = TextsRow
{ _language :: a
, _space0 :: a
, _name :: a
, _space1 :: a
, _abbreviation :: a
, _space2 :: a
, _disambig :: a
} deriving (Functor, Foldable, Traversable)
langWidgetId :: Widget.Id -> LangId -> Widget.Id
langWidgetId parentId lang =
parentId `Widget.joinId` [encodeS lang]
nameId :: Widget.Id -> Widget.Id
nameId = (`Widget.joinId` ["name"])
hspace :: _ => m (TextWidget f)
hspace = Spacer.stdHSpace <&> Widget.fromView <&> WithTextPos 0
hspaceOf :: Widget.R -> TextWidget f
hspaceOf w = Spacer.makeHorizontal w & Widget.fromView & WithTextPos 0
textsRow ::
_ =>
m (TextWidget f) ->
m (TextWidget f) ->
m (TextWidget f) ->
m (TextWidget f) ->
m (TextsRow (Aligned (Widget f)))
textsRow lang name abbrev disambig =
TextsRow lang hspace name hspace abbrev hspace disambig
& sequenceA
<&> Lens.mapped %~ Align.fromWithTextPos 0
makeLangRow ::
_ =>
Widget.Id -> (LangId -> TextsInLang -> o ()) -> LangId -> TextsInLang ->
m (TextsRow (Aligned (Widget o)))
makeLangRow parentId setName lang langNames =
textsRow
(makeLanguageTitle langId lang & info)
(makeFocusableTagNameEdit (nameId langId) nameProp)
(mkProp Tag.abbreviation & makeFocusableTagNameEdit (mkId "abbr"))
(mkProp Tag.disambiguationText & makeFocusableTagNameEdit (mkId "disamb"))
where
mkId suffix = langId `Widget.joinId` [suffix]
langId = langWidgetId parentId lang
nameProp =
setName lang . (\x -> langNames & Tag.name .~ x)
& Property (langNames ^. Tag.name)
mkProp l =
setName lang .
(\x -> langNames & Lens.cloneLens l .~ if x == "" then Nothing else Just x)
& Property (langNames ^. Lens.cloneLens l . Lens._Just)
makeMissingLangRow ::
_ =>
Widget.Id -> (LangId -> TextsInLang -> o ()) -> LangId ->
m (TextsRow (Aligned (Widget o)))
makeMissingLangRow parentId setName lang =
textsRow
(makeLanguageTitle langId lang & info)
(makeFocusableTagNameEdit (nameId langId) nameProp)
(pure Element.empty)
(pure Element.empty)
where
langId = langWidgetId parentId lang
nameProp =
setName lang . (\x -> TextsInLang x Nothing Nothing)
& Property ""
makeLangsTable ::
(MonadReader env m, _) =>
Widget.Id -> Map LangId TextsInLang ->
(LangId -> TextsInLang -> o ()) -> m (Widget o)
makeLangsTable myId tagTexts setName =
do
lang <- Lens.view has
let currentLang =
case tagTexts ^. Lens.at lang of
Nothing -> makeMissingLangRow myId setName lang
Just cur -> makeLangRow myId setName lang cur
let editOtherLangs =
tagTexts ^@.. Lens.itraversed
& filter ((/= lang) . fst)
<&> uncurry (makeLangRow myId setName)
Grid.make <*>
sequence
(heading : currentLang : editOtherLangs)
<&> snd
where
-- the type of Styled.label is RankN, so duplicate a bit of
-- code to avoid complicating too much here
toWidget = fmap Widget.fromView
heading =
textsRow
(label MomentuTexts.language <&> toWidget)
(label Texts.name <&> toWidget)
(label Texts.abbreviation <&> toWidget)
(label Texts.disambiguationText <&> toWidget)
& info
data SymType = NoSymbol | UniversalSymbol | DirectionalSymbol
deriving Eq
makeSymbol ::
_ => Widget.Id -> Property o Tag.Symbol -> m (TextWidget o, TextWidget o)
makeSymbol myId symProp =
case symProp ^. pVal of
Tag.NoSymbol ->
flip (,) Element.empty <$> makeDropDownList NoSymbol (toSym "" "")
Tag.UniversalSymbol text ->
(,)
<$> makeDropDownList UniversalSymbol (toSym text text)
<*> nameEdit (Property text (set . Tag.UniversalSymbol)) "universal"
Tag.DirectionalSymbol (Tag.DirOp ltr rtl) ->
(,)
<$> makeDropDownList DirectionalSymbol (toSym ltr rtl)
<*>
( (label Texts.leftToRightSymbol & info <&> fmap Widget.fromView)
/|/ hspace /|/ nameEdit (Property ltr (`setDirectional` rtl)) "ltr"
/|/ hspace /|/ info (label Texts.rightToLeftSymbol)
/|/ hspace /|/ nameEdit (Property rtl (setDirectional ltr)) "rtl"
)
where
set = void . Property.set symProp
setDirectional ltr rtl = Tag.DirOp ltr rtl & Tag.DirectionalSymbol & set
toSym _ _ NoSymbol = Tag.NoSymbol
toSym "" rtl UniversalSymbol = Tag.UniversalSymbol rtl
toSym ltr _ UniversalSymbol = Tag.UniversalSymbol ltr
toSym ltr rtl DirectionalSymbol = Tag.DirectionalSymbol (Tag.DirOp ltr rtl)
mkId suffix = myId `Widget.joinId` [suffix]
nameEdit prop = makeSymbolNameEdit prop . mkId
focusableLabel l suffix =
TextView.makeFocusable <*> Lens.view (has . l) ?? mkId suffix
makeDropDownList curType toTagSym =
do
noSymLabel <- focusableLabel Texts.noSymbol "nosym"
uniLabel <- focusableLabel Texts.symbol "unisym"
dirLabel <- focusableLabel Texts.directionalSymbol "dirsym"
defConf <- DropDownList.defaultConfig <*> Lens.view (has . Texts.symbolType)
DropDownList.make ?? Property curType (set . toTagSym)
?? [ (NoSymbol, noSymLabel)
, (UniversalSymbol, uniLabel)
, (DirectionalSymbol, dirLabel)
]
?? defConf ?? mkId "symType"
& withColor TextColors.actionTextColor
parseInt :: Text -> Maybe Int
parseInt newText
| newText /= Text.strip newText = Nothing
| newText == "" = Just 0
| otherwise = tryParse newText
makeIntEdit :: _ => Widget.Id -> Property o Int -> m (TextWidget o)
makeIntEdit myId prop =
do
text <-
M.readWidgetState myId
<&> (^? Lens._Just . Lens.filtered ((== Just prevVal) . parseInt))
<&> fromMaybe prevValStr
TextEdit.make ?? TextEdit.Modes "0" "0" ?? text ?? myId
<&> Align.tValue . Widget.eventMapMaker . Lens.mapped %~
-- Avoid taking keys that don't belong to us,
-- so weakerEvents with them will work.
E.filter (Lens.has Lens._Just . parseInt . fst)
<&> Align.tValue . Widget.updates %~
\(newText, eventRes) ->
eventRes <> GuiState.updateWidgetState myId newText
<$ (parseInt newText & parseAssert & Property.set prop)
where
parseAssert = error "parsing int failed" & fromMaybe
prevVal = Property.value prop
prevValStr = show prevVal & Text.pack
makeOrderEdit :: _ => Widget.Id -> Property o Int -> m (TextWidget o)
makeOrderEdit tagPaneId prop =
info (label Texts.order) /|/ hspace /|/
makeIntEdit orderEditId prop
where
orderEditId = tagPaneId `Widget.joinId` ["tagOrder"]
make :: _ => Sugar.TagPane o -> M.WidgetId -> m (Widget o)
make tagPane myId =
Lens.view has
>>= \lang ->
addValFrame <*>
do
(symbol, nextLine) <- makeSymbol myId symbolProp
langsTable <-
makeLangsTable myId
(tagPane ^. Sugar.tpTagData . Tag.tagTexts) (tagPane ^. Sugar.tpSetTexts)
orderEdit <- makeOrderEdit myId orderProp
let totalWidth =
max (nextLine ^. Element.width) (langsTable ^. Element.width)
let gap = totalWidth - (symbol ^. Element.width + orderEdit ^. Element.width)
pure langsTable
/-/ (hbox ?? [symbol, hspaceOf gap, orderEdit] <&> (^. Align.tValue))
/-/ pure (nextLine ^. Align.tValue)
& local (Element.animIdPrefix .~ Widget.toAnimId myId)
& GuiState.assignCursor myId (nameId (langWidgetId myId lang))
where
prop lens setterLens =
Property
(tagPane ^. Sugar.tpTagData . lens)
(tagPane ^. setterLens)
orderProp = prop Tag.tagOrder Sugar.tpSetOrder
symbolProp = prop Tag.tagSymbol Sugar.tpSetSymbol
| Peaker/lamdu | src/Lamdu/GUI/TagPane.hs | gpl-3.0 | 12,550 | 0 | 20 | 3,527 | 3,632 | 1,913 | 1,719 | -1 | -1 |
module Stats (Stats,
initialStats,
levelInStats,
linesInStats,
scoreInStats,
addLinesToStats) where
data Stats = Stats { linesCount :: Int, score :: Int }
initialStats :: Stats
initialStats = Stats 0 0
levelInStats :: Stats -> Int
levelInStats stats = div (linesCount stats) 5
linesInStats :: Stats -> Int
linesInStats = linesCount
scoreInStats :: Stats -> Int
scoreInStats = score
addLinesToStats :: Int -> Stats -> Stats
addLinesToStats n (Stats l s) = Stats (l + n) (s + scoreForLines n)
where scoreForLines :: Int -> Int
scoreForLines 0 = 0
scoreForLines 1 = 100
scoreForLines 2 = 200
scoreForLines 3 = 400
scoreForLines 4 = 800
scoreForLines _ = error "Wrong number of lines" | pavelfatin/haskell-blocks | src/Stats.hs | gpl-3.0 | 820 | 0 | 8 | 251 | 231 | 125 | 106 | 24 | 6 |
-- Author: Viacheslav Lotsmanov
-- License: GPLv3 https://raw.githubusercontent.com/unclechu/xlib-keys-hack/master/LICENSE
module Utils.Sugar (spec) where
import "hspec" Test.Hspec (Spec, describe, it, shouldBe)
-- local imports
import "xlib-keys-hack" Utils.Sugar
( (.>), (|?|), (?)
, applyIf, applyUnless
)
spec :: Spec
spec = do
describe "Piping operators" $
it "(.>) flipped version of composition operator" $ do
(subtract 6 . (*2) . subtract 2) 15 `shouldBe` (20 :: Int)
(subtract 2 .> (*2) .> subtract 6) 15 `shouldBe` (20 :: Int)
describe "Boolean operators" $ do
it "(|?|) flipped `bool` function (A value if True or B value if False)" $ do
(10 |?| 20) True `shouldBe` (10 :: Int)
(10 |?| 20) False `shouldBe` (20 :: Int)
it "(?) if-condition operator" $ do
(True ? 10 $ 20) `shouldBe` (10 :: Int)
(False ? 10 $ 20) `shouldBe` (20 :: Int)
it "(?) if-condition operator chaining" $
(False ? 10 $ False ? 20 $ True ? 30 $ 40) `shouldBe` (30 :: Int)
describe "Function applying helpers" $ do
it "`applyIf` applies given function if value passes predicate\
\ or returns original value" $ do
applyIf (+5) True 10 `shouldBe` (15 :: Int)
applyIf (+5) False 10 `shouldBe` (10 :: Int)
it "`applyUnless` applies given function if value doesn't pass predicate\
\ or returns original value" $ do
applyUnless (+5) False 10 `shouldBe` (15 :: Int)
applyUnless (+5) True 10 `shouldBe` (10 :: Int)
| unclechu/xlib-keys-hack | test/Utils/Sugar.hs | gpl-3.0 | 1,530 | 0 | 18 | 372 | 483 | 267 | 216 | -1 | -1 |
module KRPCHS.RemoteTech
( Target(..)
, Antenna
, Comms
, antenna
, antennaStream
, antennaStreamReq
, getAntennaHasConnection
, getAntennaHasConnectionStream
, getAntennaHasConnectionStreamReq
, getAntennaPart
, getAntennaPartStream
, getAntennaPartStreamReq
, getAntennaTarget
, getAntennaTargetStream
, getAntennaTargetStreamReq
, getAntennaTargetBody
, getAntennaTargetBodyStream
, getAntennaTargetBodyStreamReq
, getAntennaTargetGroundStation
, getAntennaTargetGroundStationStream
, getAntennaTargetGroundStationStreamReq
, getAntennaTargetVessel
, getAntennaTargetVesselStream
, getAntennaTargetVesselStreamReq
, setAntennaTarget
, setAntennaTargetBody
, setAntennaTargetGroundStation
, setAntennaTargetVessel
, comms
, commsStream
, commsStreamReq
, commsSignalDelayToVessel
, commsSignalDelayToVesselStream
, commsSignalDelayToVesselStreamReq
, getCommsAntennas
, getCommsAntennasStream
, getCommsAntennasStreamReq
, getCommsHasConnection
, getCommsHasConnectionStream
, getCommsHasConnectionStreamReq
, getCommsHasConnectionToGroundStation
, getCommsHasConnectionToGroundStationStream
, getCommsHasConnectionToGroundStationStreamReq
, getCommsHasFlightComputer
, getCommsHasFlightComputerStream
, getCommsHasFlightComputerStreamReq
, getCommsHasLocalControl
, getCommsHasLocalControlStream
, getCommsHasLocalControlStreamReq
, getCommsSignalDelay
, getCommsSignalDelayStream
, getCommsSignalDelayStreamReq
, getCommsSignalDelayToGroundStation
, getCommsSignalDelayToGroundStationStream
, getCommsSignalDelayToGroundStationStreamReq
, getCommsVessel
, getCommsVesselStream
, getCommsVesselStreamReq
, getAvailable
, getAvailableStream
, getAvailableStreamReq
, getGroundStations
, getGroundStationsStream
, getGroundStationsStreamReq
) where
import qualified Data.Text
import qualified KRPCHS.SpaceCenter
import KRPCHS.Internal.Requests
import KRPCHS.Internal.SerializeUtils
{-
- A RemoteTech antenna. Obtained by calling <see cref="M:RemoteTech.Comms.Antennas" /> or <see cref="M:RemoteTech.Antenna" />.
-}
newtype Antenna = Antenna { antennaId :: Int }
deriving (Show, Eq, Ord)
instance PbSerializable Antenna where
encodePb = encodePb . antennaId
decodePb b = Antenna <$> decodePb b
instance KRPCResponseExtractable Antenna
{-
- Communications for a vessel.
-}
newtype Comms = Comms { commsId :: Int }
deriving (Show, Eq, Ord)
instance PbSerializable Comms where
encodePb = encodePb . commsId
decodePb b = Comms <$> decodePb b
instance KRPCResponseExtractable Comms
{-
- The type of object an antenna is targetting.
- See <see cref="M:RemoteTech.Antenna.Target" />.
-}
data Target
= Target'ActiveVessel
| Target'CelestialBody
| Target'GroundStation
| Target'Vessel
| Target'None
deriving (Show, Eq, Ord, Enum)
instance PbSerializable Target where
encodePb = encodePb . fromEnum
decodePb b = toEnum <$> decodePb b
instance KRPCResponseExtractable Target
{-
- Get the antenna object for a particular part.
-}
antenna :: KRPCHS.SpaceCenter.Part -> RPCContext (KRPCHS.RemoteTech.Antenna)
antenna partArg = do
let r = makeRequest "RemoteTech" "Antenna" [makeArgument 0 partArg]
res <- sendRequest r
processResponse res
antennaStreamReq :: KRPCHS.SpaceCenter.Part -> KRPCStreamReq (KRPCHS.RemoteTech.Antenna)
antennaStreamReq partArg =
let req = makeRequest "RemoteTech" "Antenna" [makeArgument 0 partArg]
in makeStream req
antennaStream :: KRPCHS.SpaceCenter.Part -> RPCContext (KRPCStream (KRPCHS.RemoteTech.Antenna))
antennaStream partArg = requestStream $ antennaStreamReq partArg
{-
- Whether the antenna has a connection.
-}
getAntennaHasConnection :: KRPCHS.RemoteTech.Antenna -> RPCContext (Bool)
getAntennaHasConnection thisArg = do
let r = makeRequest "RemoteTech" "Antenna_get_HasConnection" [makeArgument 0 thisArg]
res <- sendRequest r
processResponse res
getAntennaHasConnectionStreamReq :: KRPCHS.RemoteTech.Antenna -> KRPCStreamReq (Bool)
getAntennaHasConnectionStreamReq thisArg =
let req = makeRequest "RemoteTech" "Antenna_get_HasConnection" [makeArgument 0 thisArg]
in makeStream req
getAntennaHasConnectionStream :: KRPCHS.RemoteTech.Antenna -> RPCContext (KRPCStream (Bool))
getAntennaHasConnectionStream thisArg = requestStream $ getAntennaHasConnectionStreamReq thisArg
{-
- Get the part containing this antenna.
-}
getAntennaPart :: KRPCHS.RemoteTech.Antenna -> RPCContext (KRPCHS.SpaceCenter.Part)
getAntennaPart thisArg = do
let r = makeRequest "RemoteTech" "Antenna_get_Part" [makeArgument 0 thisArg]
res <- sendRequest r
processResponse res
getAntennaPartStreamReq :: KRPCHS.RemoteTech.Antenna -> KRPCStreamReq (KRPCHS.SpaceCenter.Part)
getAntennaPartStreamReq thisArg =
let req = makeRequest "RemoteTech" "Antenna_get_Part" [makeArgument 0 thisArg]
in makeStream req
getAntennaPartStream :: KRPCHS.RemoteTech.Antenna -> RPCContext (KRPCStream (KRPCHS.SpaceCenter.Part))
getAntennaPartStream thisArg = requestStream $ getAntennaPartStreamReq thisArg
{-
- The object that the antenna is targetting.
- This property can be used to set the target to <see cref="M:RemoteTech.Target.None" /> or <see cref="M:RemoteTech.Target.ActiveVessel" />.
- To set the target to a celestial body, ground station or vessel see <see cref="M:RemoteTech.Antenna.TargetBody" />,
- <see cref="M:RemoteTech.Antenna.TargetGroundStation" /> and <see cref="M:RemoteTech.Antenna.TargetVessel" />.
-}
getAntennaTarget :: KRPCHS.RemoteTech.Antenna -> RPCContext (KRPCHS.RemoteTech.Target)
getAntennaTarget thisArg = do
let r = makeRequest "RemoteTech" "Antenna_get_Target" [makeArgument 0 thisArg]
res <- sendRequest r
processResponse res
getAntennaTargetStreamReq :: KRPCHS.RemoteTech.Antenna -> KRPCStreamReq (KRPCHS.RemoteTech.Target)
getAntennaTargetStreamReq thisArg =
let req = makeRequest "RemoteTech" "Antenna_get_Target" [makeArgument 0 thisArg]
in makeStream req
getAntennaTargetStream :: KRPCHS.RemoteTech.Antenna -> RPCContext (KRPCStream (KRPCHS.RemoteTech.Target))
getAntennaTargetStream thisArg = requestStream $ getAntennaTargetStreamReq thisArg
{-
- The celestial body the antenna is targetting.
-}
getAntennaTargetBody :: KRPCHS.RemoteTech.Antenna -> RPCContext (KRPCHS.SpaceCenter.CelestialBody)
getAntennaTargetBody thisArg = do
let r = makeRequest "RemoteTech" "Antenna_get_TargetBody" [makeArgument 0 thisArg]
res <- sendRequest r
processResponse res
getAntennaTargetBodyStreamReq :: KRPCHS.RemoteTech.Antenna -> KRPCStreamReq (KRPCHS.SpaceCenter.CelestialBody)
getAntennaTargetBodyStreamReq thisArg =
let req = makeRequest "RemoteTech" "Antenna_get_TargetBody" [makeArgument 0 thisArg]
in makeStream req
getAntennaTargetBodyStream :: KRPCHS.RemoteTech.Antenna -> RPCContext (KRPCStream (KRPCHS.SpaceCenter.CelestialBody))
getAntennaTargetBodyStream thisArg = requestStream $ getAntennaTargetBodyStreamReq thisArg
{-
- The ground station the antenna is targetting.
-}
getAntennaTargetGroundStation :: KRPCHS.RemoteTech.Antenna -> RPCContext (Data.Text.Text)
getAntennaTargetGroundStation thisArg = do
let r = makeRequest "RemoteTech" "Antenna_get_TargetGroundStation" [makeArgument 0 thisArg]
res <- sendRequest r
processResponse res
getAntennaTargetGroundStationStreamReq :: KRPCHS.RemoteTech.Antenna -> KRPCStreamReq (Data.Text.Text)
getAntennaTargetGroundStationStreamReq thisArg =
let req = makeRequest "RemoteTech" "Antenna_get_TargetGroundStation" [makeArgument 0 thisArg]
in makeStream req
getAntennaTargetGroundStationStream :: KRPCHS.RemoteTech.Antenna -> RPCContext (KRPCStream (Data.Text.Text))
getAntennaTargetGroundStationStream thisArg = requestStream $ getAntennaTargetGroundStationStreamReq thisArg
{-
- The vessel the antenna is targetting.
-}
getAntennaTargetVessel :: KRPCHS.RemoteTech.Antenna -> RPCContext (KRPCHS.SpaceCenter.Vessel)
getAntennaTargetVessel thisArg = do
let r = makeRequest "RemoteTech" "Antenna_get_TargetVessel" [makeArgument 0 thisArg]
res <- sendRequest r
processResponse res
getAntennaTargetVesselStreamReq :: KRPCHS.RemoteTech.Antenna -> KRPCStreamReq (KRPCHS.SpaceCenter.Vessel)
getAntennaTargetVesselStreamReq thisArg =
let req = makeRequest "RemoteTech" "Antenna_get_TargetVessel" [makeArgument 0 thisArg]
in makeStream req
getAntennaTargetVesselStream :: KRPCHS.RemoteTech.Antenna -> RPCContext (KRPCStream (KRPCHS.SpaceCenter.Vessel))
getAntennaTargetVesselStream thisArg = requestStream $ getAntennaTargetVesselStreamReq thisArg
{-
- The object that the antenna is targetting.
- This property can be used to set the target to <see cref="M:RemoteTech.Target.None" /> or <see cref="M:RemoteTech.Target.ActiveVessel" />.
- To set the target to a celestial body, ground station or vessel see <see cref="M:RemoteTech.Antenna.TargetBody" />,
- <see cref="M:RemoteTech.Antenna.TargetGroundStation" /> and <see cref="M:RemoteTech.Antenna.TargetVessel" />.
-}
setAntennaTarget :: KRPCHS.RemoteTech.Antenna -> KRPCHS.RemoteTech.Target -> RPCContext ()
setAntennaTarget thisArg valueArg = do
let r = makeRequest "RemoteTech" "Antenna_set_Target" [makeArgument 0 thisArg, makeArgument 1 valueArg]
res <- sendRequest r
processResponse res
{-
- The celestial body the antenna is targetting.
-}
setAntennaTargetBody :: KRPCHS.RemoteTech.Antenna -> KRPCHS.SpaceCenter.CelestialBody -> RPCContext ()
setAntennaTargetBody thisArg valueArg = do
let r = makeRequest "RemoteTech" "Antenna_set_TargetBody" [makeArgument 0 thisArg, makeArgument 1 valueArg]
res <- sendRequest r
processResponse res
{-
- The ground station the antenna is targetting.
-}
setAntennaTargetGroundStation :: KRPCHS.RemoteTech.Antenna -> Data.Text.Text -> RPCContext ()
setAntennaTargetGroundStation thisArg valueArg = do
let r = makeRequest "RemoteTech" "Antenna_set_TargetGroundStation" [makeArgument 0 thisArg, makeArgument 1 valueArg]
res <- sendRequest r
processResponse res
{-
- The vessel the antenna is targetting.
-}
setAntennaTargetVessel :: KRPCHS.RemoteTech.Antenna -> KRPCHS.SpaceCenter.Vessel -> RPCContext ()
setAntennaTargetVessel thisArg valueArg = do
let r = makeRequest "RemoteTech" "Antenna_set_TargetVessel" [makeArgument 0 thisArg, makeArgument 1 valueArg]
res <- sendRequest r
processResponse res
{-
- Get a communications object, representing the communication capability of a particular vessel.
-}
comms :: KRPCHS.SpaceCenter.Vessel -> RPCContext (KRPCHS.RemoteTech.Comms)
comms vesselArg = do
let r = makeRequest "RemoteTech" "Comms" [makeArgument 0 vesselArg]
res <- sendRequest r
processResponse res
commsStreamReq :: KRPCHS.SpaceCenter.Vessel -> KRPCStreamReq (KRPCHS.RemoteTech.Comms)
commsStreamReq vesselArg =
let req = makeRequest "RemoteTech" "Comms" [makeArgument 0 vesselArg]
in makeStream req
commsStream :: KRPCHS.SpaceCenter.Vessel -> RPCContext (KRPCStream (KRPCHS.RemoteTech.Comms))
commsStream vesselArg = requestStream $ commsStreamReq vesselArg
{-
- The signal delay between the this vessel and another vessel, in seconds.<param name="other">
-}
commsSignalDelayToVessel :: KRPCHS.RemoteTech.Comms -> KRPCHS.SpaceCenter.Vessel -> RPCContext (Double)
commsSignalDelayToVessel thisArg otherArg = do
let r = makeRequest "RemoteTech" "Comms_SignalDelayToVessel" [makeArgument 0 thisArg, makeArgument 1 otherArg]
res <- sendRequest r
processResponse res
commsSignalDelayToVesselStreamReq :: KRPCHS.RemoteTech.Comms -> KRPCHS.SpaceCenter.Vessel -> KRPCStreamReq (Double)
commsSignalDelayToVesselStreamReq thisArg otherArg =
let req = makeRequest "RemoteTech" "Comms_SignalDelayToVessel" [makeArgument 0 thisArg, makeArgument 1 otherArg]
in makeStream req
commsSignalDelayToVesselStream :: KRPCHS.RemoteTech.Comms -> KRPCHS.SpaceCenter.Vessel -> RPCContext (KRPCStream (Double))
commsSignalDelayToVesselStream thisArg otherArg = requestStream $ commsSignalDelayToVesselStreamReq thisArg otherArg
{-
- The antennas for this vessel.
-}
getCommsAntennas :: KRPCHS.RemoteTech.Comms -> RPCContext ([KRPCHS.RemoteTech.Antenna])
getCommsAntennas thisArg = do
let r = makeRequest "RemoteTech" "Comms_get_Antennas" [makeArgument 0 thisArg]
res <- sendRequest r
processResponse res
getCommsAntennasStreamReq :: KRPCHS.RemoteTech.Comms -> KRPCStreamReq ([KRPCHS.RemoteTech.Antenna])
getCommsAntennasStreamReq thisArg =
let req = makeRequest "RemoteTech" "Comms_get_Antennas" [makeArgument 0 thisArg]
in makeStream req
getCommsAntennasStream :: KRPCHS.RemoteTech.Comms -> RPCContext (KRPCStream ([KRPCHS.RemoteTech.Antenna]))
getCommsAntennasStream thisArg = requestStream $ getCommsAntennasStreamReq thisArg
{-
- Whether the vessel has any connection.
-}
getCommsHasConnection :: KRPCHS.RemoteTech.Comms -> RPCContext (Bool)
getCommsHasConnection thisArg = do
let r = makeRequest "RemoteTech" "Comms_get_HasConnection" [makeArgument 0 thisArg]
res <- sendRequest r
processResponse res
getCommsHasConnectionStreamReq :: KRPCHS.RemoteTech.Comms -> KRPCStreamReq (Bool)
getCommsHasConnectionStreamReq thisArg =
let req = makeRequest "RemoteTech" "Comms_get_HasConnection" [makeArgument 0 thisArg]
in makeStream req
getCommsHasConnectionStream :: KRPCHS.RemoteTech.Comms -> RPCContext (KRPCStream (Bool))
getCommsHasConnectionStream thisArg = requestStream $ getCommsHasConnectionStreamReq thisArg
{-
- Whether the vessel has a connection to a ground station.
-}
getCommsHasConnectionToGroundStation :: KRPCHS.RemoteTech.Comms -> RPCContext (Bool)
getCommsHasConnectionToGroundStation thisArg = do
let r = makeRequest "RemoteTech" "Comms_get_HasConnectionToGroundStation" [makeArgument 0 thisArg]
res <- sendRequest r
processResponse res
getCommsHasConnectionToGroundStationStreamReq :: KRPCHS.RemoteTech.Comms -> KRPCStreamReq (Bool)
getCommsHasConnectionToGroundStationStreamReq thisArg =
let req = makeRequest "RemoteTech" "Comms_get_HasConnectionToGroundStation" [makeArgument 0 thisArg]
in makeStream req
getCommsHasConnectionToGroundStationStream :: KRPCHS.RemoteTech.Comms -> RPCContext (KRPCStream (Bool))
getCommsHasConnectionToGroundStationStream thisArg = requestStream $ getCommsHasConnectionToGroundStationStreamReq thisArg
{-
- Whether the vessel has a flight computer on board.
-}
getCommsHasFlightComputer :: KRPCHS.RemoteTech.Comms -> RPCContext (Bool)
getCommsHasFlightComputer thisArg = do
let r = makeRequest "RemoteTech" "Comms_get_HasFlightComputer" [makeArgument 0 thisArg]
res <- sendRequest r
processResponse res
getCommsHasFlightComputerStreamReq :: KRPCHS.RemoteTech.Comms -> KRPCStreamReq (Bool)
getCommsHasFlightComputerStreamReq thisArg =
let req = makeRequest "RemoteTech" "Comms_get_HasFlightComputer" [makeArgument 0 thisArg]
in makeStream req
getCommsHasFlightComputerStream :: KRPCHS.RemoteTech.Comms -> RPCContext (KRPCStream (Bool))
getCommsHasFlightComputerStream thisArg = requestStream $ getCommsHasFlightComputerStreamReq thisArg
{-
- Whether the vessel can be controlled locally.
-}
getCommsHasLocalControl :: KRPCHS.RemoteTech.Comms -> RPCContext (Bool)
getCommsHasLocalControl thisArg = do
let r = makeRequest "RemoteTech" "Comms_get_HasLocalControl" [makeArgument 0 thisArg]
res <- sendRequest r
processResponse res
getCommsHasLocalControlStreamReq :: KRPCHS.RemoteTech.Comms -> KRPCStreamReq (Bool)
getCommsHasLocalControlStreamReq thisArg =
let req = makeRequest "RemoteTech" "Comms_get_HasLocalControl" [makeArgument 0 thisArg]
in makeStream req
getCommsHasLocalControlStream :: KRPCHS.RemoteTech.Comms -> RPCContext (KRPCStream (Bool))
getCommsHasLocalControlStream thisArg = requestStream $ getCommsHasLocalControlStreamReq thisArg
{-
- The shortest signal delay to the vessel, in seconds.
-}
getCommsSignalDelay :: KRPCHS.RemoteTech.Comms -> RPCContext (Double)
getCommsSignalDelay thisArg = do
let r = makeRequest "RemoteTech" "Comms_get_SignalDelay" [makeArgument 0 thisArg]
res <- sendRequest r
processResponse res
getCommsSignalDelayStreamReq :: KRPCHS.RemoteTech.Comms -> KRPCStreamReq (Double)
getCommsSignalDelayStreamReq thisArg =
let req = makeRequest "RemoteTech" "Comms_get_SignalDelay" [makeArgument 0 thisArg]
in makeStream req
getCommsSignalDelayStream :: KRPCHS.RemoteTech.Comms -> RPCContext (KRPCStream (Double))
getCommsSignalDelayStream thisArg = requestStream $ getCommsSignalDelayStreamReq thisArg
{-
- The signal delay between the vessel and the closest ground station, in seconds.
-}
getCommsSignalDelayToGroundStation :: KRPCHS.RemoteTech.Comms -> RPCContext (Double)
getCommsSignalDelayToGroundStation thisArg = do
let r = makeRequest "RemoteTech" "Comms_get_SignalDelayToGroundStation" [makeArgument 0 thisArg]
res <- sendRequest r
processResponse res
getCommsSignalDelayToGroundStationStreamReq :: KRPCHS.RemoteTech.Comms -> KRPCStreamReq (Double)
getCommsSignalDelayToGroundStationStreamReq thisArg =
let req = makeRequest "RemoteTech" "Comms_get_SignalDelayToGroundStation" [makeArgument 0 thisArg]
in makeStream req
getCommsSignalDelayToGroundStationStream :: KRPCHS.RemoteTech.Comms -> RPCContext (KRPCStream (Double))
getCommsSignalDelayToGroundStationStream thisArg = requestStream $ getCommsSignalDelayToGroundStationStreamReq thisArg
{-
- Get the vessel.
-}
getCommsVessel :: KRPCHS.RemoteTech.Comms -> RPCContext (KRPCHS.SpaceCenter.Vessel)
getCommsVessel thisArg = do
let r = makeRequest "RemoteTech" "Comms_get_Vessel" [makeArgument 0 thisArg]
res <- sendRequest r
processResponse res
getCommsVesselStreamReq :: KRPCHS.RemoteTech.Comms -> KRPCStreamReq (KRPCHS.SpaceCenter.Vessel)
getCommsVesselStreamReq thisArg =
let req = makeRequest "RemoteTech" "Comms_get_Vessel" [makeArgument 0 thisArg]
in makeStream req
getCommsVesselStream :: KRPCHS.RemoteTech.Comms -> RPCContext (KRPCStream (KRPCHS.SpaceCenter.Vessel))
getCommsVesselStream thisArg = requestStream $ getCommsVesselStreamReq thisArg
{-
- Whether RemoteTech is installed.
-}
getAvailable :: RPCContext (Bool)
getAvailable = do
let r = makeRequest "RemoteTech" "get_Available" []
res <- sendRequest r
processResponse res
getAvailableStreamReq :: KRPCStreamReq (Bool)
getAvailableStreamReq =
let req = makeRequest "RemoteTech" "get_Available" []
in makeStream req
getAvailableStream :: RPCContext (KRPCStream (Bool))
getAvailableStream = requestStream $ getAvailableStreamReq
{-
- The names of the ground stations.
-}
getGroundStations :: RPCContext ([Data.Text.Text])
getGroundStations = do
let r = makeRequest "RemoteTech" "get_GroundStations" []
res <- sendRequest r
processResponse res
getGroundStationsStreamReq :: KRPCStreamReq ([Data.Text.Text])
getGroundStationsStreamReq =
let req = makeRequest "RemoteTech" "get_GroundStations" []
in makeStream req
getGroundStationsStream :: RPCContext (KRPCStream ([Data.Text.Text]))
getGroundStationsStream = requestStream $ getGroundStationsStreamReq
| Cahu/krpc-hs | src/KRPCHS/RemoteTech.hs | gpl-3.0 | 19,170 | 0 | 12 | 2,534 | 3,896 | 1,981 | 1,915 | 321 | 1 |
{-# LANGUAGE DataKinds #-}
{-# LANGUAGE DeriveDataTypeable #-}
{-# LANGUAGE DeriveGeneric #-}
{-# LANGUAGE FlexibleInstances #-}
{-# LANGUAGE NoImplicitPrelude #-}
{-# LANGUAGE OverloadedStrings #-}
{-# LANGUAGE RecordWildCards #-}
{-# LANGUAGE TypeFamilies #-}
{-# LANGUAGE TypeOperators #-}
{-# OPTIONS_GHC -fno-warn-duplicate-exports #-}
{-# OPTIONS_GHC -fno-warn-unused-binds #-}
{-# OPTIONS_GHC -fno-warn-unused-imports #-}
-- |
-- Module : Network.Google.Resource.AndroidEnterprise.ManagedConfigurationssettings.List
-- Copyright : (c) 2015-2016 Brendan Hay
-- License : Mozilla Public License, v. 2.0.
-- Maintainer : Brendan Hay <[email protected]>
-- Stability : auto-generated
-- Portability : non-portable (GHC extensions)
--
-- Lists all the managed configurations settings for the specified app.
--
-- /See:/ <https://developers.google.com/android/work/play/emm-api Google Play EMM API Reference> for @androidenterprise.managedconfigurationssettings.list@.
module Network.Google.Resource.AndroidEnterprise.ManagedConfigurationssettings.List
(
-- * REST Resource
ManagedConfigurationssettingsListResource
-- * Creating a Request
, managedConfigurationssettingsList
, ManagedConfigurationssettingsList
-- * Request Lenses
, mclXgafv
, mclUploadProtocol
, mclEnterpriseId
, mclAccessToken
, mclUploadType
, mclProductId
, mclCallback
) where
import Network.Google.AndroidEnterprise.Types
import Network.Google.Prelude
-- | A resource alias for @androidenterprise.managedconfigurationssettings.list@ method which the
-- 'ManagedConfigurationssettingsList' request conforms to.
type ManagedConfigurationssettingsListResource =
"androidenterprise" :>
"v1" :>
"enterprises" :>
Capture "enterpriseId" Text :>
"products" :>
Capture "productId" Text :>
"managedConfigurationsSettings" :>
QueryParam "$.xgafv" Xgafv :>
QueryParam "upload_protocol" Text :>
QueryParam "access_token" Text :>
QueryParam "uploadType" Text :>
QueryParam "callback" Text :>
QueryParam "alt" AltJSON :>
Get '[JSON]
ManagedConfigurationsSettingsListResponse
-- | Lists all the managed configurations settings for the specified app.
--
-- /See:/ 'managedConfigurationssettingsList' smart constructor.
data ManagedConfigurationssettingsList =
ManagedConfigurationssettingsList'
{ _mclXgafv :: !(Maybe Xgafv)
, _mclUploadProtocol :: !(Maybe Text)
, _mclEnterpriseId :: !Text
, _mclAccessToken :: !(Maybe Text)
, _mclUploadType :: !(Maybe Text)
, _mclProductId :: !Text
, _mclCallback :: !(Maybe Text)
}
deriving (Eq, Show, Data, Typeable, Generic)
-- | Creates a value of 'ManagedConfigurationssettingsList' with the minimum fields required to make a request.
--
-- Use one of the following lenses to modify other fields as desired:
--
-- * 'mclXgafv'
--
-- * 'mclUploadProtocol'
--
-- * 'mclEnterpriseId'
--
-- * 'mclAccessToken'
--
-- * 'mclUploadType'
--
-- * 'mclProductId'
--
-- * 'mclCallback'
managedConfigurationssettingsList
:: Text -- ^ 'mclEnterpriseId'
-> Text -- ^ 'mclProductId'
-> ManagedConfigurationssettingsList
managedConfigurationssettingsList pMclEnterpriseId_ pMclProductId_ =
ManagedConfigurationssettingsList'
{ _mclXgafv = Nothing
, _mclUploadProtocol = Nothing
, _mclEnterpriseId = pMclEnterpriseId_
, _mclAccessToken = Nothing
, _mclUploadType = Nothing
, _mclProductId = pMclProductId_
, _mclCallback = Nothing
}
-- | V1 error format.
mclXgafv :: Lens' ManagedConfigurationssettingsList (Maybe Xgafv)
mclXgafv = lens _mclXgafv (\ s a -> s{_mclXgafv = a})
-- | Upload protocol for media (e.g. \"raw\", \"multipart\").
mclUploadProtocol :: Lens' ManagedConfigurationssettingsList (Maybe Text)
mclUploadProtocol
= lens _mclUploadProtocol
(\ s a -> s{_mclUploadProtocol = a})
-- | The ID of the enterprise.
mclEnterpriseId :: Lens' ManagedConfigurationssettingsList Text
mclEnterpriseId
= lens _mclEnterpriseId
(\ s a -> s{_mclEnterpriseId = a})
-- | OAuth access token.
mclAccessToken :: Lens' ManagedConfigurationssettingsList (Maybe Text)
mclAccessToken
= lens _mclAccessToken
(\ s a -> s{_mclAccessToken = a})
-- | Legacy upload protocol for media (e.g. \"media\", \"multipart\").
mclUploadType :: Lens' ManagedConfigurationssettingsList (Maybe Text)
mclUploadType
= lens _mclUploadType
(\ s a -> s{_mclUploadType = a})
-- | The ID of the product for which the managed configurations settings
-- applies to.
mclProductId :: Lens' ManagedConfigurationssettingsList Text
mclProductId
= lens _mclProductId (\ s a -> s{_mclProductId = a})
-- | JSONP
mclCallback :: Lens' ManagedConfigurationssettingsList (Maybe Text)
mclCallback
= lens _mclCallback (\ s a -> s{_mclCallback = a})
instance GoogleRequest
ManagedConfigurationssettingsList
where
type Rs ManagedConfigurationssettingsList =
ManagedConfigurationsSettingsListResponse
type Scopes ManagedConfigurationssettingsList =
'["https://www.googleapis.com/auth/androidenterprise"]
requestClient ManagedConfigurationssettingsList'{..}
= go _mclEnterpriseId _mclProductId _mclXgafv
_mclUploadProtocol
_mclAccessToken
_mclUploadType
_mclCallback
(Just AltJSON)
androidEnterpriseService
where go
= buildClient
(Proxy ::
Proxy ManagedConfigurationssettingsListResource)
mempty
| brendanhay/gogol | gogol-android-enterprise/gen/Network/Google/Resource/AndroidEnterprise/ManagedConfigurationssettings/List.hs | mpl-2.0 | 5,926 | 0 | 20 | 1,385 | 786 | 457 | 329 | 122 | 1 |
-- Copyright (C) 2016-2017 Red Hat, Inc.
--
-- This library is free software; you can redistribute it and/or
-- modify it under the terms of the GNU Lesser General Public
-- License as published by the Free Software Foundation; either
-- version 2.1 of the License, or (at your option) any later version.
--
-- This library is distributed in the hope that it will be useful,
-- but WITHOUT ANY WARRANTY; without even the implied warranty of
-- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
-- Lesser General Public License for more details.
--
-- You should have received a copy of the GNU Lesser General Public
-- License along with this library; if not, see <http://www.gnu.org/licenses/>.
{-# LANGUAGE OverloadedStrings #-}
module BDCS.Packages(filesInPackage,
findPackage,
insertPackageName)
where
import Control.Monad.IO.Class(MonadIO)
import qualified Data.Text as T
import Database.Esqueleto
import BDCS.DB
import BDCS.KeyValue(insertKeyValue)
filesInPackage :: MonadIO m => T.Text -> SqlPersistT m [T.Text]
filesInPackage name = do
results <- select $ from $ \(files `InnerJoin` key_val `InnerJoin` file_key_values) -> do
on $ key_val ^. KeyValId ==. file_key_values ^. FileKeyValuesKey_val_id &&.
file_key_values ^. FileKeyValuesFile_id ==. files ^. FilesId
where_ $ key_val ^. KeyValKey_value ==. val "packageName" &&.
key_val ^. KeyValVal_value ==. val name
return $ files ^. FilesPath
return $ map unValue results
insertPackageName :: MonadIO m => T.Text -> SqlPersistT m (Key KeyVal)
insertPackageName packageName =
findPackage packageName `orDo` insertKeyValue "packageName" packageName Nothing
findPackage :: MonadIO m => T.Text -> SqlPersistT m (Maybe (Key KeyVal))
findPackage name = firstResult $
select $ from $ \pkg -> do
where_ $ pkg ^. KeyValKey_value ==. val "packageName" &&.
pkg ^. KeyValVal_value ==. val name
limit 1
return $ pkg ^. KeyValId
| dashea/bdcs | importer/BDCS/Packages.hs | lgpl-2.1 | 2,097 | 0 | 19 | 489 | 397 | 210 | 187 | 28 | 1 |
module Main where
import Points (convexHull, center)
import Point (Point)
import Support (shouldEql)
p00 = (0, 0)
p10 = (1, 0)
p10' = (-1, 0)
p01 = (0, 1)
p01' = (0, -1)
p11 = (1, 1)
p11' = (-1, -1)
p20 = (2, 0)
p02 = (0, 2)
main :: IO ()
main = print $ foldl1 (++) [
center [p00] `shouldEql` p00,
center [p10', p10] `shouldEql` p00,
center [p01', p01] `shouldEql` p00,
center [p01', p01, p10] `shouldEql` (0.5, 0),
center [
(-1, 1), p11,
p11', (1, -1)
] `shouldEql` p00,
center [
(-1, 1), p11,
p20,
p11', (1, -1)
] `shouldEql` (0.5, 0),
convexHull [] `shouldEql` [],
convexHull [p00] `shouldEql` [],
convexHull [p00, p01] `shouldEql` [],
convexHull [p00, p01, p01'] `shouldEql` [],
convexHull [p01, p01', p00] `shouldEql` [],
convexHull [p01', p00, p01] `shouldEql` [],
convexHull [p00, p10, p10', p20] `shouldEql` [],
convexHull [p00, p01, p01', p02] `shouldEql` [],
convexHull [p00, p01, p10] `shouldEql` [p01, p10, p00],
convexHull [p01, p10, p00] `shouldEql` [p01, p10, p00],
convexHull [p10, p00, p01] `shouldEql` [p01, p10, p00],
convexHull [
(-1, 1), ( 1, 1),
(-1, -1), ( 1, -1)
] `shouldEql` [(1, 1), (1, -1), (-1, -1), (-1, 1)],
convexHull [
(-1, 1), ( 1, 1),
( 2, 0),
(-1, -1), ( 1, -1)
] `shouldEql` [(1, 1), (2, 0), (1, -1), (-1, -1), (-1, 1)],
convexHull [
(-1, 1), ( 0, 1), ( 1, 1),
(-0.5, 0), ( 0, 0), (1.5, 0),
(-1, -1), ( 0, -1)
] `shouldEql` [(1, 1), (1.5, 0), (0, -1), (-1, -1), (-1, 1)],
convexHull [
(-2, 2), ( 2, 2),
(-1, 1), ( 0, 1), ( 1, 1),
(-1, 0), ( 0, 0), ( 1, 0),
(-1, -1), ( 0, -1), ( 1, -1),
(-2, -2), ( 2, -2)
] `shouldEql` [(2, 2), (2, -2), (-2, -2), (-2, 2)],
convexHull [
(-2, 2),
(-1, 1),
( 0, 0),
( 1, -1),
( 2, -2)
] `shouldEql` [],
convexHull [
( 0, 2),
(-1, 1), ( 1, 1),
(-1, -1), ( 1, -1),
( 0, -2)
] `shouldEql` [(0, 2), (1, 1), (1, -1), (0, -2), (-1, -1), (-1, 1)],
convexHull [
(-2, 0), ( 0, -0.5), ( 2, 0),
(-1, -1), ( 1, -1),
( 0, -2)
] `shouldEql` [(2, 0), (0, -2), (-2, 0)],
convexHull [
(-2, 0), ( 0, -0.5), ( 2, 0),
(-1, -1.5), ( 1, -1.5),
( 0, -2)
] `shouldEql` [(2, 0), (1, -1.5), (0, -2), (-1, -1.5), (-2, 0)],
convexHull [
( 0, 1),
(-2, 0), ( 0, -0.5), ( 2, 0),
(-1, -1), ( 1, -1),
( 0, -2),
( 0, -3)
] `shouldEql` [(0, 1), (2, 0), (0, -3), (-2, 0)],
"done"
]
| jhnns/haskell-experiments | test/PointsSpec.hs | unlicense | 3,532 | 0 | 12 | 1,762 | 1,743 | 1,096 | 647 | 91 | 1 |
findKey :: (Eq k) => k -> [(k, v)] -> Maybe v
findKey _ [] = Nothing
findKey key ((k, v):xs)
| key == k = Just v
| otherwise = findKey key xs
findKey' :: (Eq k) => k -> [(k, v)] -> Maybe v
findKey' _ [] = Nothing
findKey' key ((k, v):xs) = case key == k of
True -> Just v
False -> findKey' key xs
| EricYT/Haskell | src/chapter-3-2.hs | apache-2.0 | 366 | 1 | 9 | 139 | 192 | 100 | 92 | 10 | 2 |
module FractalFlame.Variation.Types.VTransform where
import FractalFlame.Point.Types.CartesianPoint
import FractalFlame.Variation.Types.VarP
-- | Variation function (use FractalFlame.Variation.runVariation to build a VarP parameter)
type VTransform = VarP -> CartesianPoint
| anthezium/fractal_flame_renderer_haskell | FractalFlame/Variation/Types/VTransform.hs | bsd-2-clause | 277 | 0 | 5 | 27 | 33 | 23 | 10 | 4 | 0 |
{-# LANGUAGE MultiParamTypeClasses #-}
{-# LANGUAGE TypeFamilies #-}
{-# LANGUAGE DataKinds #-}
module Data.Interface.Module.Entity where
import Data.Interface.Change
import Data.Interface.Name
import Data.Interface.Type
import Data.Interface.Type.Diff
-- | A top-level exported entity in a module, without a name.
data Entity
= LocalValue ValueDecl -- ^ a value-namespace declaration
| LocalType TypeDecl -- ^ a type-namespace declaration
| ReExport ModuleName Namespace -- ^ module and namespace of declaration
deriving (Show, Eq, Ord)
data EntityDiff
= LocalValueDiff ValueDeclDiff -- ^ a value-namespace diff
| LocalTypeDiff TypeDeclDiff -- ^ a type-namespace diff
| EntityDiff (Change Entity) -- ^ none of the above
deriving (Show, Eq, Ord)
instance ToChange Entity EntityDiff where
toChange ediff = case ediff of
LocalValueDiff vd -> LocalValue <$> toChange vd
LocalTypeDiff td -> LocalType <$> toChange td
EntityDiff c -> c
instance Diff Entity EntityDiff where
noDiff e = case e of
LocalValue vd -> LocalValueDiff (noDiff vd)
LocalType td -> LocalTypeDiff (noDiff td)
ReExport{} -> EntityDiff (NoChange e)
diff a b = case (a,b) of
(LocalValue vd0, LocalValue vd1) -> LocalValueDiff (diff vd0 vd1)
(LocalType td0, LocalType td1) -> LocalTypeDiff (diff td0 td1)
_ -> EntityDiff (diff a b)
-- * ValueDecl
data ValueDecl = ValueDecl
{ vdType :: Type
, vdInfo :: ValueDeclInfo
} deriving (Show, Eq, Ord)
data ValueDeclInfo
= Identifier
| PatternSyn
| DataCon [DataField]
deriving (Show, Eq, Ord)
type DataField = Named ()
type instance Space ValueDecl = 'Values
instance HasNamespace ValueDecl where
namespace _ = Values
instance TraverseNames ValueDecl where
traverseNames f (ValueDecl t i) =
ValueDecl <$> traverseNames f t <*> traverseNames f i
instance TraverseNames ValueDeclInfo where
traverseNames f vdi = case vdi of
DataCon fields -> DataCon <$> traverse (traverseNames f) fields
_ -> pure vdi
data ValueDeclDiff = ValueDeclDiff
{ vdTypeDiff :: TypeDiff
, vdInfoDiff :: Change ValueDeclInfo
} deriving (Show, Eq, Ord)
instance ToChange ValueDecl ValueDeclDiff where
toChange (ValueDeclDiff t i) = ValueDecl <$> toChange t <*> toChange i
instance Diff ValueDecl ValueDeclDiff where
diff (ValueDecl ta ia) (ValueDecl tb ib) =
ValueDeclDiff (diff ta tb) (diff ia ib)
-- * TypeDecl
data TypeDecl = TypeDecl
{ tdKind :: Kind
, tdInfo :: TypeDeclInfo
} deriving (Show, Eq, Ord)
data TypeDeclInfo
= DataType DataConList -- ^ data/newtype
| TypeSyn String -- ^ type synonym (TODO)
| TypeClass -- ^ type class (TODO)
deriving (Show, Eq, Ord)
{- TypeDecl notes:
- TypeSyn contains its definition
(this is only a String for now, but will have to include
first-class type information)
TODO:
- type/data families
-}
type instance Space TypeDecl = 'Types
instance HasNamespace TypeDecl where
namespace _ = Types
instance TraverseNames TypeDecl where
traverseNames f (TypeDecl k i) =
TypeDecl <$> traverseNames f k <*> traverseNames f i
instance TraverseNames TypeDeclInfo where
traverseNames f tdi = case tdi of
DataType dcons -> DataType <$> traverseNames f dcons
_ -> pure tdi
-- | Data constructors for an algebraic type, or `Abstract` when the data
-- constructors are hidden.
data DataConList = Abstract | DataConList [RawName]
deriving (Show, Eq, Ord)
instance TraverseNames DataConList where
traverseNames f dcons = case dcons of
Abstract -> pure Abstract
DataConList ns -> DataConList <$> traverse (traverseNames f) ns
data TypeDeclDiff = TypeDeclDiff
{ tdKindDiff :: Change Kind
, tdInfoDiff :: Change TypeDeclInfo
} deriving (Show, Eq, Ord)
instance ToChange TypeDecl TypeDeclDiff where
toChange (TypeDeclDiff t i) = TypeDecl <$> toChange t <*> toChange i
instance Diff TypeDecl TypeDeclDiff where
diff (TypeDecl ka ia) (TypeDecl kb ib) =
TypeDeclDiff (diff ka kb) (diff ia ib)
| cdxr/haskell-interface | src/Data/Interface/Module/Entity.hs | bsd-3-clause | 4,273 | 0 | 12 | 1,064 | 1,118 | 583 | 535 | 95 | 0 |
{-# LANGUAGE MultiWayIf #-}
{-# LANGUAGE DeriveAnyClass #-}
{-# LANGUAGE DeriveGeneric #-}
{-# LANGUAGE FlexibleInstances #-}
{-# LANGUAGE MultiParamTypeClasses #-}
{-# LANGUAGE TemplateHaskell #-}
{-|
The 'Message' is a single displayed event in a Channel. All Messages
have a date/time, and messages that represent posts to the channel
have a (hash) ID, and displayable text, along with other attributes.
All Messages are sorted chronologically. There is no assumption that
the server date/time is synchronized with the local date/time, so all
of the Message ordering uses the server's date/time.
The mattermost-api retrieves a 'Post' from the server, briefly encodes
the useful portions of that as a 'ClientPost' object and then converts
it to a 'Message' inserting this result it into the collection of
Messages associated with a Channel. The PostID of the message
uniquely identifies that message and can be used to interact with the
server for subsequent operations relative to that message's 'Post'.
The date/time associated with these messages is generated by the
server.
There are also "messages" generated directly by the Matterhorn client
which can be used to display additional, client-related information to
the user. Examples of these client messages are: date boundaries, the
"new messages" marker, errors from invoking the browser, etc. These
client-generated messages will have a date/time although it is locally
generated (usually by relation to an associated Post).
Most other Matterhorn operations primarily are concerned with
user-posted messages (@case mMessageId of Just _@ or @case mType of CP
_@), but others will include client-generated messages (@case mMessageId
of Nothing@ or @case mType of C _@).
--}
module Matterhorn.Types.Messages
( -- * Message and operations on a single Message
Message(..)
, isDeletable, isReplyable, isReactable, isEditable, isReplyTo, isGap, isFlaggable
, isPinnable, isEmote, isJoinLeave, isTransition, isNewMessagesTransition
, mText, mUser, mDate, mType, mPending, mDeleted, mPinned
, mAttachments, mInReplyToMsg, mMessageId, mReactions, mFlagged
, mOriginalPost, mChannelId, mMarkdownSource
, isBotMessage
, MessageType(..)
, MessageId(..)
, ThreadState(..)
, MentionedUser(..)
, isPostMessage
, messagePostId
, messageIdPostId
, UserRef(..)
, ReplyState(..)
, clientMessageToMessage
, clientPostToMessage
, clientPostReactionUserIds
, newMessageOfType
-- * Message Collections
, Messages
, ChronologicalMessages
, RetrogradeMessages
, MessageOps (..)
, noMessages
, messagesLength
, filterMessages
, reverseMessages
, unreverseMessages
, splitMessages
, splitDirSeqOn
, chronologicalMsgsWithThreadStates
, retrogradeMsgsWithThreadStates
, findMessage
, getRelMessageId
, messagesHead
, messagesDrop
, getNextMessage
, getPrevMessage
, getNextMessageId
, getPrevMessageId
, getNextPostId
, getPrevPostId
, getEarliestPostMsg
, getLatestPostMsg
, getEarliestSelectableMessage
, getLatestSelectableMessage
, findLatestUserMessage
-- * Operations on any Message type
, messagesAfter
, removeMatchesFromSubset
, withFirstMessage
, msgURLs
, LinkTarget(..)
, LinkChoice(LinkChoice, _linkTarget)
, linkUser
, linkTarget
, linkTime
, linkLabel
)
where
import Prelude ()
import Matterhorn.Prelude
import Control.Monad
import qualified Data.Foldable as F
import Data.Hashable ( Hashable )
import qualified Data.Map.Strict as Map
import Data.Sequence as Seq
import qualified Data.Set as S
import Data.Tuple
import Data.UUID ( UUID )
import GHC.Generics ( Generic )
import Lens.Micro.Platform ( makeLenses )
import Network.Mattermost.Types ( ChannelId, PostId, Post
, ServerTime, UserId, FileId
)
import Matterhorn.Types.DirectionalSeq
import Matterhorn.Types.Posts
import Matterhorn.Types.RichText
-- | The state of a message's thread context.
data ThreadState =
NoThread
-- ^ The message is not in a thread at all.
| InThreadShowParent
-- ^ The message is in a thread, and the thread's root message
-- (parent) should be displayed above this message.
| InThread
-- ^ The message is in a thread but the thread's root message should
-- not be displayed above this message.
deriving (Show, Eq)
-- ----------------------------------------------------------------------
-- * Messages
data MessageId = MessagePostId PostId
| MessageUUID UUID
deriving (Eq, Read, Ord, Show, Generic, Hashable)
messageIdPostId :: MessageId -> Maybe PostId
messageIdPostId (MessagePostId p) = Just p
messageIdPostId _ = Nothing
-- | A 'Message' is any message we might want to render, either from
-- Mattermost itself or from a client-internal source.
data Message = Message
{ _mText :: Blocks
, _mMarkdownSource :: Text
, _mUser :: UserRef
, _mDate :: ServerTime
, _mType :: MessageType
, _mPending :: Bool
, _mDeleted :: Bool
, _mAttachments :: Seq Attachment
, _mInReplyToMsg :: ReplyState
, _mMessageId :: Maybe MessageId
, _mReactions :: Map.Map Text (S.Set UserId)
, _mOriginalPost :: Maybe Post
, _mFlagged :: Bool
, _mPinned :: Bool
, _mChannelId :: Maybe ChannelId
} deriving (Show)
isPostMessage :: Message -> Bool
isPostMessage m =
isJust (_mMessageId m >>= messageIdPostId)
messagePostId :: Message -> Maybe PostId
messagePostId m = do
mId <- _mMessageId m
messageIdPostId mId
isDeletable :: Message -> Bool
isDeletable m =
isJust (messagePostId m) &&
case _mType m of
CP NormalPost -> True
CP Emote -> True
_ -> False
isFlaggable :: Message -> Bool
isFlaggable = isJust . messagePostId
isPinnable :: Message -> Bool
isPinnable = isJust . messagePostId
isReplyable :: Message -> Bool
isReplyable m =
isJust (messagePostId m) &&
case _mType m of
CP NormalPost -> True
CP Emote -> True
_ -> False
isReactable :: Message -> Bool
isReactable m =
isJust (messagePostId m) &&
case _mType m of
CP NormalPost -> True
CP Emote -> True
_ -> False
isEditable :: Message -> Bool
isEditable m =
isJust (messagePostId m) &&
case _mType m of
CP NormalPost -> True
CP Emote -> True
_ -> False
isReplyTo :: PostId -> Message -> Bool
isReplyTo expectedParentId m =
case _mInReplyToMsg m of
NotAReply -> False
InReplyTo actualParentId -> actualParentId == expectedParentId
isGap :: Message -> Bool
isGap m = case _mType m of
C UnknownGapBefore -> True
C UnknownGapAfter -> True
_ -> False
isTransition :: Message -> Bool
isTransition m = case _mType m of
C DateTransition -> True
C NewMessagesTransition -> True
_ -> False
isNewMessagesTransition :: Message -> Bool
isNewMessagesTransition m = case _mType m of
C NewMessagesTransition -> True
_ -> False
isEmote :: Message -> Bool
isEmote m = case _mType m of
CP Emote -> True
_ -> False
isJoinLeave :: Message -> Bool
isJoinLeave m = case _mType m of
CP Join -> True
CP Leave -> True
_ -> False
-- | A 'Message' is the representation we use for storage and
-- rendering, so it must be able to represent either a
-- post from Mattermost or an internal message. This represents
-- the union of both kinds of post types.
data MessageType = C ClientMessageType
| CP ClientPostType
deriving (Show)
-- | There may be no user (usually an internal message), a reference to
-- a user (by Id), or the server may have supplied a specific username
-- (often associated with bots). The boolean flag indicates whether the
-- user reference is for a message from a bot.
data UserRef = NoUser | UserI Bool UserId | UserOverride Bool Text
deriving (Eq, Show, Ord)
isBotMessage :: Message -> Bool
isBotMessage m =
case _mUser m of
UserI bot _ -> bot
UserOverride bot _ -> bot
NoUser -> False
-- | The 'ReplyState' of a message represents whether a message
-- is a reply, and if so, to what message
data ReplyState =
NotAReply
| InReplyTo PostId
deriving (Show, Eq)
data LinkTarget =
LinkURL URL
| LinkFileId FileId
| LinkPermalink TeamURLName PostId
deriving (Eq, Show, Ord)
-- | This type represents links to things in the 'open links' view.
data LinkChoice =
LinkChoice { _linkTime :: ServerTime
, _linkUser :: UserRef
, _linkLabel :: Maybe Inlines
, _linkTarget :: LinkTarget
} deriving (Eq, Show)
makeLenses ''LinkChoice
-- | Convert a 'ClientMessage' to a 'Message'. A 'ClientMessage' is
-- one that was generated by the Matterhorn client and which the
-- server knows nothing about. For example, an error message
-- associated with passing a link to the local browser.
clientMessageToMessage :: ClientMessage -> Message
clientMessageToMessage cm = Message
{ _mText = parseMarkdown Nothing (cm^.cmText)
, _mMarkdownSource = cm^.cmText
, _mUser = NoUser
, _mDate = cm^.cmDate
, _mType = C $ cm^.cmType
, _mPending = False
, _mDeleted = False
, _mAttachments = Seq.empty
, _mInReplyToMsg = NotAReply
, _mMessageId = Nothing
, _mReactions = Map.empty
, _mOriginalPost = Nothing
, _mFlagged = False
, _mPinned = False
, _mChannelId = Nothing
}
data MentionedUser =
UsernameMention Text
| UserIdMention UserId
deriving (Eq, Show, Ord)
clientPostReactionUserIds :: ClientPost -> S.Set UserId
clientPostReactionUserIds cp =
S.unions $ F.toList $ cp^.cpReactions
-- | Builds a message from a ClientPost and also returns the set of
-- usernames mentioned in the text of the message.
clientPostToMessage :: ClientPost -> (Message, S.Set MentionedUser)
clientPostToMessage cp = (m, mentions)
where
mentions =
S.fromList $
(UsernameMention <$> (F.toList $ findUsernames $ cp^.cpText)) <>
(UserIdMention <$> (F.toList $ clientPostReactionUserIds cp))
m = Message { _mText = cp^.cpText
, _mMarkdownSource = cp^.cpMarkdownSource
, _mUser =
case cp^.cpUserOverride of
Just n | cp^.cpType == NormalPost -> UserOverride (cp^.cpFromWebhook) n
_ -> maybe NoUser (UserI (cp^.cpFromWebhook)) $ cp^.cpUser
, _mDate = cp^.cpDate
, _mType = CP $ cp^.cpType
, _mPending = cp^.cpPending
, _mDeleted = cp^.cpDeleted
, _mAttachments = cp^.cpAttachments
, _mInReplyToMsg =
case cp^.cpInReplyToPost of
Nothing -> NotAReply
Just pId -> InReplyTo pId
, _mMessageId = Just $ MessagePostId $ cp^.cpPostId
, _mReactions = cp^.cpReactions
, _mOriginalPost = Just $ cp^.cpOriginalPost
, _mFlagged = False
, _mPinned = cp^.cpPinned
, _mChannelId = Just $ cp^.cpChannelId
}
newMessageOfType :: Text -> MessageType -> ServerTime -> Message
newMessageOfType text typ d = Message
{ _mText = parseMarkdown Nothing text
, _mMarkdownSource = text
, _mUser = NoUser
, _mDate = d
, _mType = typ
, _mPending = False
, _mDeleted = False
, _mAttachments = Seq.empty
, _mInReplyToMsg = NotAReply
, _mMessageId = Nothing
, _mReactions = Map.empty
, _mOriginalPost = Nothing
, _mFlagged = False
, _mPinned = False
, _mChannelId = Nothing
}
-- ** 'Message' Lenses
makeLenses ''Message
-- ----------------------------------------------------------------------
-- * Message Collections
-- | A wrapper for an ordered, unique list of 'Message' values.
--
-- This type has (and promises) the following instances: Show,
-- Functor, Monoid, Foldable, Traversable
type ChronologicalMessages = DirectionalSeq Chronological Message
type Messages = ChronologicalMessages
-- | There are also cases where the list of 'Message' values are kept
-- in reverse order (most recent -> oldest); these cases are
-- represented by the `RetrogradeMessages` type.
type RetrogradeMessages = DirectionalSeq Retrograde Message
-- ** Common operations on Messages
filterMessages :: SeqDirection seq
=> (a -> Bool)
-> DirectionalSeq seq a
-> DirectionalSeq seq a
filterMessages f = onDirectedSeq (Seq.filter f)
class MessageOps a where
-- | addMessage inserts a date in proper chronological order, with
-- the following extra functionality:
-- * no duplication (by PostId)
-- * no duplication (adjacent UnknownGap entries)
addMessage :: Message -> a -> a
instance MessageOps ChronologicalMessages where
addMessage m ml =
case viewr (dseq ml) of
EmptyR -> DSeq $ singleton m
_ :> l ->
case compare (m^.mDate) (l^.mDate) of
GT -> DSeq $ dseq ml |> m
EQ -> if m^.mMessageId == l^.mMessageId && isJust (m^.mMessageId)
then ml
else dirDateInsert m ml
LT -> dirDateInsert m ml
dirDateInsert :: Message -> ChronologicalMessages -> ChronologicalMessages
dirDateInsert m = onDirectedSeq $ finalize . foldr insAfter initial
where initial = (Just m, mempty)
insAfter c (Nothing, l) = (Nothing, c <| l)
insAfter c (Just n, l) =
case compare (n^.mDate) (c^.mDate) of
GT -> (Nothing, c <| (n <| l))
EQ -> if n^.mMessageId == c^.mMessageId && isJust (c^.mMessageId)
then (Nothing, c <| l)
else (Just n, c <| l)
LT -> (Just n, c <| l)
finalize (Just n, l) = n <| l
finalize (_, l) = l
noMessages :: Messages
noMessages = DSeq mempty
messagesLength :: DirectionalSeq seq a -> Int
messagesLength (DSeq ms) = Seq.length ms
-- | Reverse the order of the messages
reverseMessages :: Messages -> RetrogradeMessages
reverseMessages = DSeq . Seq.reverse . dseq
-- | Unreverse the order of the messages
unreverseMessages :: RetrogradeMessages -> Messages
unreverseMessages = DSeq . Seq.reverse . dseq
splitDirSeqOn :: SeqDirection d
=> (a -> Bool)
-> DirectionalSeq d a
-> (Maybe a, (DirectionalSeq (ReverseDirection d) a,
DirectionalSeq d a))
splitDirSeqOn f msgs =
let (removed, remaining) = dirSeqBreakl f msgs
devomer = DSeq $ Seq.reverse $ dseq removed
in (withDirSeqHead id remaining, (devomer, onDirectedSeq (Seq.drop 1) remaining))
-- ----------------------------------------------------------------------
-- * Operations on Posted Messages
-- | Searches for the specified MessageId and returns a tuple where the
-- first element is the Message associated with the MessageId (if it
-- exists), and the second element is another tuple: the first element
-- of the second is all the messages from the beginning of the list to
-- the message just before the MessageId message (or all messages if not
-- found) *in reverse order*, and the second element of the second are
-- all the messages that follow the found message (none if the message
-- was never found) in *forward* order.
splitMessages :: Maybe MessageId
-> DirectionalSeq Chronological (Message, ThreadState)
-> (Maybe (Message, ThreadState),
( DirectionalSeq Retrograde (Message, ThreadState),
DirectionalSeq Chronological (Message, ThreadState)))
splitMessages mid msgs = splitDirSeqOn (\(m, _) -> isJust mid && m^.mMessageId == mid) msgs
-- | Given a message and its chronological predecessor, return
-- the thread state of the specified message with respect to its
-- predecessor.
threadStateFor :: Message
-- ^ The message whose state is to be obtained.
-> Message
-- ^ The message's predecessor.
-> ThreadState
threadStateFor msg prev = case msg^.mInReplyToMsg of
InReplyTo rootId ->
if | (prev^.mMessageId) == Just (MessagePostId rootId) ->
InThread
| prev^.mInReplyToMsg == msg^.mInReplyToMsg ->
InThread
| otherwise ->
InThreadShowParent
_ -> NoThread
retrogradeMsgsWithThreadStates :: RetrogradeMessages -> DirectionalSeq Retrograde (Message, ThreadState)
retrogradeMsgsWithThreadStates msgs = DSeq $ checkAdjacentMessages (dseq msgs)
where
getMessagePredecessor ms =
let visiblePredMsg m = not (isTransition m || m^.mDeleted) in
case Seq.viewl ms of
prev Seq.:< rest ->
if visiblePredMsg prev
then Just prev
else getMessagePredecessor rest
Seq.EmptyL -> Nothing
checkAdjacentMessages s = case Seq.viewl s of
Seq.EmptyL -> mempty
m Seq.:< t ->
let new_m = case getMessagePredecessor t of
Just prev -> (m, threadStateFor m prev)
Nothing -> case m^.mInReplyToMsg of
InReplyTo _ -> (m, InThreadShowParent)
_ -> (m, NoThread)
in new_m Seq.<| checkAdjacentMessages t
chronologicalMsgsWithThreadStates :: Messages -> DirectionalSeq Chronological (Message, ThreadState)
chronologicalMsgsWithThreadStates msgs = DSeq $ checkAdjacentMessages (dseq msgs)
where
getMessagePredecessor ms =
let visiblePredMsg m = not (isTransition m || m^.mDeleted) in
case Seq.viewr ms of
rest Seq.:> prev ->
if visiblePredMsg prev
then Just prev
else getMessagePredecessor rest
Seq.EmptyR -> Nothing
checkAdjacentMessages s = case Seq.viewr s of
Seq.EmptyR -> mempty
t Seq.:> m ->
let new_m = case getMessagePredecessor t of
Just prev -> (m, threadStateFor m prev)
Nothing -> case m^.mInReplyToMsg of
InReplyTo _ -> (m, InThreadShowParent)
_ -> (m, NoThread)
in checkAdjacentMessages t Seq.|> new_m
-- | findMessage searches for a specific message as identified by the
-- PostId. The search starts from the most recent messages because
-- that is the most likely place the message will occur.
findMessage :: MessageId -> Messages -> Maybe Message
findMessage mid msgs =
findIndexR (\m -> m^.mMessageId == Just mid) (dseq msgs)
>>= Just . Seq.index (dseq msgs)
-- | Look forward for the first Message with an ID that follows the
-- specified Id and return it. If no input Id supplied, get the
-- latest (most recent chronologically) Message in the input set.
getNextMessage :: Maybe MessageId -> Messages -> Maybe Message
getNextMessage = getRelMessageId
-- | Look backward for the first Message with an ID that follows the
-- specified MessageId and return it. If no input MessageId supplied,
-- get the latest (most recent chronologically) Message in the input
-- set.
getPrevMessage :: Maybe MessageId -> Messages -> Maybe Message
getPrevMessage mId = getRelMessageId mId . reverseMessages
messagesHead :: (SeqDirection seq) => DirectionalSeq seq a -> Maybe a
messagesHead = withDirSeqHead id
messagesDrop :: (SeqDirection seq) => Int -> DirectionalSeq seq a -> DirectionalSeq seq a
messagesDrop i = onDirectedSeq (Seq.drop i)
-- | Look forward for the first Message with an ID that follows the
-- specified MessageId and return that found Message's ID; if no input
-- MessageId is specified, return the latest (most recent
-- chronologically) MessageId (if any) in the input set.
getNextMessageId :: Maybe MessageId -> Messages -> Maybe MessageId
getNextMessageId mId = _mMessageId <=< getNextMessage mId
-- | Look backwards for the first Message with an ID that comes before
-- the specified MessageId and return that found Message's ID; if no
-- input MessageId is specified, return the latest (most recent
-- chronologically) MessageId (if any) in the input set.
getPrevMessageId :: Maybe MessageId -> Messages -> Maybe MessageId
getPrevMessageId mId = _mMessageId <=< getPrevMessage mId
-- | Look forward for the first Message with an ID that follows the
-- specified PostId and return that found Message's PostID; if no
-- input PostId is specified, return the latest (most recent
-- chronologically) PostId (if any) in the input set.
getNextPostId :: Maybe PostId -> Messages -> Maybe PostId
getNextPostId pid = messagePostId <=< getNextMessage (MessagePostId <$> pid)
-- | Look backwards for the first Post with an ID that comes before
-- the specified PostId.
getPrevPostId :: Maybe PostId -> Messages -> Maybe PostId
getPrevPostId pid = messagePostId <=< getPrevMessage (MessagePostId <$> pid)
getRelMessageId :: SeqDirection dir =>
Maybe MessageId
-> DirectionalSeq dir Message
-> Maybe Message
getRelMessageId mId =
let isMId = const ((==) mId . _mMessageId) <$> mId
in getRelMessage isMId
-- | Internal worker function to return a different user message in
-- relation to either the latest point or a specific message.
getRelMessage :: SeqDirection dir =>
Maybe (Message -> Bool)
-> DirectionalSeq dir Message
-> Maybe Message
getRelMessage matcher msgs =
let after = case matcher of
Just matchFun -> case splitDirSeqOn matchFun msgs of
(_, (_, ms)) -> ms
Nothing -> msgs
in withDirSeqHead id $ filterMessages validSelectableMessage after
-- | Find the most recent message that is a Post (as opposed to a
-- local message) (if any).
getLatestPostMsg :: Messages -> Maybe Message
getLatestPostMsg msgs =
case viewr $ dropWhileR (not . validUserMessage) (dseq msgs) of
EmptyR -> Nothing
_ :> m -> Just m
-- | Find the oldest message that is a message with an ID.
getEarliestSelectableMessage :: Messages -> Maybe Message
getEarliestSelectableMessage msgs =
case viewl $ dropWhileL (not . validSelectableMessage) (dseq msgs) of
EmptyL -> Nothing
m :< _ -> Just m
-- | Find the most recent message that is a message with an ID.
getLatestSelectableMessage :: Messages -> Maybe Message
getLatestSelectableMessage msgs =
case viewr $ dropWhileR (not . validSelectableMessage) (dseq msgs) of
EmptyR -> Nothing
_ :> m -> Just m
-- | Find the earliest message that is a Post (as opposed to a
-- local message) (if any).
getEarliestPostMsg :: Messages -> Maybe Message
getEarliestPostMsg msgs =
case viewl $ dropWhileL (not . validUserMessage) (dseq msgs) of
EmptyL -> Nothing
m :< _ -> Just m
-- | Find the most recent message that is a message posted by a user
-- that matches the test (if any), skipping local client messages and
-- any user event that is not a message (i.e. find a normal message or
-- an emote).
findLatestUserMessage :: (Message -> Bool) -> Messages -> Maybe Message
findLatestUserMessage f ml =
case viewr $ dropWhileR (\m -> not (validUserMessage m && f m)) $ dseq ml of
EmptyR -> Nothing
_ :> m -> Just m
validUserMessage :: Message -> Bool
validUserMessage m =
not (m^.mDeleted) && case m^.mMessageId of
Just (MessagePostId _) -> True
_ -> False
validSelectableMessage :: Message -> Bool
validSelectableMessage m = (not $ m^.mDeleted) && (isJust $ m^.mMessageId)
-- ----------------------------------------------------------------------
-- * Operations on any Message type
-- | Return all messages that were posted after the specified date/time.
messagesAfter :: ServerTime -> Messages -> Messages
messagesAfter viewTime = onDirectedSeq $ takeWhileR (\m -> m^.mDate > viewTime)
-- | Removes any Messages (all types) for which the predicate is true
-- from the specified subset of messages (identified by a starting and
-- ending MessageId, inclusive) and returns the resulting list (from
-- start to finish, irrespective of 'firstId' and 'lastId') and the
-- list of removed items.
--
-- start | end | operates-on | (test) case
-- --------------------------------------------------------|-------------
-- Nothing | Nothing | entire list | C1
-- Nothing | Just found | start --> found] | C2
-- Nothing | Just missing | nothing [suggest invalid] | C3
-- Just found | Nothing | [found --> end | C4
-- Just found | Just found | [found --> found] | C5
-- Just found | Just missing | [found --> end | C6
-- Just missing | Nothing | nothing [suggest invalid] | C7
-- Just missing | Just found | start --> found] | C8
-- Just missing | Just missing | nothing [suggest invalid] | C9
--
-- @removeMatchesFromSubset matchPred fromId toId msgs = (remaining, removed)@
--
removeMatchesFromSubset :: (Message -> Bool) -> Maybe MessageId -> Maybe MessageId
-> Messages -> (Messages, Messages)
removeMatchesFromSubset matching firstId lastId msgs =
let knownIds = fmap (^.mMessageId) msgs
in if isNothing firstId && isNothing lastId
then swap $ dirSeqPartition matching msgs
else if isJust firstId && firstId `elem` knownIds
then onDirSeqSubset
(\m -> m^.mMessageId == firstId)
(if isJust lastId then \m -> m^.mMessageId == lastId else const False)
(swap . dirSeqPartition matching) msgs
else if isJust lastId && lastId `elem` knownIds
then onDirSeqSubset
(const True)
(\m -> m^.mMessageId == lastId)
(swap . dirSeqPartition matching) msgs
else (msgs, noMessages)
-- | Performs an operation on the first Message, returning just the
-- result of that operation, or Nothing if there were no messages.
-- Note that the message is not necessarily a posted user message.
withFirstMessage :: SeqDirection dir
=> (Message -> r)
-> DirectionalSeq dir Message
-> Maybe r
withFirstMessage = withDirSeqHead
msgURLs :: Message -> Seq LinkChoice
msgURLs msg =
let uRef = msg^.mUser
mkTarget (Right url) = LinkURL url
mkTarget (Left (tName, pId)) = LinkPermalink tName pId
mkEntry (val, text) = LinkChoice (msg^.mDate) uRef text (mkTarget val)
msgUrls = mkEntry <$> (Seq.fromList $ mconcat $ blockGetURLs <$> (F.toList $ unBlocks $ msg^.mText))
attachmentURLs = (\ a ->
LinkChoice
(msg^.mDate)
uRef
(Just $ attachmentLabel a)
(LinkFileId $ a^.attachmentFileId))
<$> (msg^.mAttachments)
attachmentLabel a =
Inlines $ Seq.fromList [ EText "attachment"
, ESpace
, ECode $ Inlines $ Seq.singleton $ EText $ a^.attachmentName
]
in msgUrls <> attachmentURLs
| matterhorn-chat/matterhorn | src/Matterhorn/Types/Messages.hs | bsd-3-clause | 28,203 | 0 | 20 | 7,952 | 5,443 | 2,939 | 2,504 | 496 | 6 |
module Seventeen where
import Data.List
permute :: Int -> [Int] -> [Int] -> [[Int]]
permute goal tried []
| goal == sum tried = [tried]
| otherwise = []
permute goal tried remaining
| goal == sum tried = [tried]
| otherwise = concatMap (permute2 goal tried) (sublists remaining)
permute2 :: Int -> [Int] -> [Int] -> [[Int]]
permute2 _ _ [] = []
permute2 goal tried (x:xs)
| x + sum tried <= goal = permute goal (tried ++ [x]) xs
| otherwise = []
sublists :: [Int] -> [[Int]]
sublists = filter (not . null) . tails
seventeen :: IO Int
seventeen = do
input <- readFile "input/17.txt"
let buckets = map read $ lines input
in return . length $ permute 150 [] buckets
| purcell/adventofcodeteam | app/Seventeen.hs | bsd-3-clause | 705 | 0 | 12 | 169 | 345 | 174 | 171 | 21 | 1 |
{-# language CPP #-}
-- | = Name
--
-- VK_AMD_shader_core_properties2 - device extension
--
-- == VK_AMD_shader_core_properties2
--
-- [__Name String__]
-- @VK_AMD_shader_core_properties2@
--
-- [__Extension Type__]
-- Device extension
--
-- [__Registered Extension Number__]
-- 228
--
-- [__Revision__]
-- 1
--
-- [__Extension and Version Dependencies__]
--
-- - Requires Vulkan 1.0
--
-- - Requires @VK_AMD_shader_core_properties@
--
-- [__Contact__]
--
-- - Matthaeus G. Chajdas
-- <https://github.com/KhronosGroup/Vulkan-Docs/issues/new?body=[VK_AMD_shader_core_properties2] @anteru%0A<<Here describe the issue or question you have about the VK_AMD_shader_core_properties2 extension>> >
--
-- == Other Extension Metadata
--
-- [__Last Modified Date__]
-- 2019-07-26
--
-- [__IP Status__]
-- No known IP claims.
--
-- [__Contributors__]
--
-- - Matthaeus G. Chajdas, AMD
--
-- - Tobias Hector, AMD
--
-- == Description
--
-- This extension exposes additional shader core properties for a target
-- physical device through the @VK_KHR_get_physical_device_properties2@
-- extension.
--
-- == New Structures
--
-- - Extending
-- 'Vulkan.Core11.Promoted_From_VK_KHR_get_physical_device_properties2.PhysicalDeviceProperties2':
--
-- - 'PhysicalDeviceShaderCoreProperties2AMD'
--
-- == New Enums
--
-- - 'ShaderCorePropertiesFlagBitsAMD'
--
-- == New Bitmasks
--
-- - 'ShaderCorePropertiesFlagsAMD'
--
-- == New Enum Constants
--
-- - 'AMD_SHADER_CORE_PROPERTIES_2_EXTENSION_NAME'
--
-- - 'AMD_SHADER_CORE_PROPERTIES_2_SPEC_VERSION'
--
-- - Extending 'Vulkan.Core10.Enums.StructureType.StructureType':
--
-- - 'Vulkan.Core10.Enums.StructureType.STRUCTURE_TYPE_PHYSICAL_DEVICE_SHADER_CORE_PROPERTIES_2_AMD'
--
-- == Examples
--
-- None.
--
-- == Version History
--
-- - Revision 1, 2019-07-26 (Matthaeus G. Chajdas)
--
-- - Initial draft.
--
-- == See Also
--
-- 'PhysicalDeviceShaderCoreProperties2AMD',
-- 'ShaderCorePropertiesFlagBitsAMD', 'ShaderCorePropertiesFlagsAMD'
--
-- == Document Notes
--
-- For more information, see the
-- <https://www.khronos.org/registry/vulkan/specs/1.3-extensions/html/vkspec.html#VK_AMD_shader_core_properties2 Vulkan Specification>
--
-- This page is a generated document. Fixes and changes should be made to
-- the generator scripts, not directly.
module Vulkan.Extensions.VK_AMD_shader_core_properties2 ( PhysicalDeviceShaderCoreProperties2AMD(..)
, ShaderCorePropertiesFlagsAMD
, ShaderCorePropertiesFlagBitsAMD(..)
, AMD_SHADER_CORE_PROPERTIES_2_SPEC_VERSION
, pattern AMD_SHADER_CORE_PROPERTIES_2_SPEC_VERSION
, AMD_SHADER_CORE_PROPERTIES_2_EXTENSION_NAME
, pattern AMD_SHADER_CORE_PROPERTIES_2_EXTENSION_NAME
) where
import Vulkan.Internal.Utils (enumReadPrec)
import Vulkan.Internal.Utils (enumShowsPrec)
import Foreign.Marshal.Alloc (allocaBytes)
import Foreign.Ptr (nullPtr)
import Foreign.Ptr (plusPtr)
import GHC.Show (showString)
import Numeric (showHex)
import Vulkan.CStruct (FromCStruct)
import Vulkan.CStruct (FromCStruct(..))
import Vulkan.CStruct (ToCStruct)
import Vulkan.CStruct (ToCStruct(..))
import Vulkan.Zero (Zero)
import Vulkan.Zero (Zero(..))
import Data.Bits (Bits)
import Data.Bits (FiniteBits)
import Data.String (IsString)
import Data.Typeable (Typeable)
import Foreign.Storable (Storable)
import Foreign.Storable (Storable(peek))
import Foreign.Storable (Storable(poke))
import qualified Foreign.Storable (Storable(..))
import GHC.Generics (Generic)
import Foreign.Ptr (Ptr)
import GHC.Read (Read(readPrec))
import GHC.Show (Show(showsPrec))
import Data.Word (Word32)
import Data.Kind (Type)
import Vulkan.Core10.FundamentalTypes (Flags)
import Vulkan.Core10.Enums.StructureType (StructureType)
import Vulkan.Core10.Enums.StructureType (StructureType(STRUCTURE_TYPE_PHYSICAL_DEVICE_SHADER_CORE_PROPERTIES_2_AMD))
-- | VkPhysicalDeviceShaderCoreProperties2AMD - Structure describing shader
-- core properties that can be supported by an implementation
--
-- = Description
--
-- If the 'PhysicalDeviceShaderCoreProperties2AMD' structure is included in
-- the @pNext@ chain of the
-- 'Vulkan.Core11.Promoted_From_VK_KHR_get_physical_device_properties2.PhysicalDeviceProperties2'
-- structure passed to
-- 'Vulkan.Core11.Promoted_From_VK_KHR_get_physical_device_properties2.getPhysicalDeviceProperties2',
-- it is filled in with each corresponding implementation-dependent
-- property.
--
-- == Valid Usage (Implicit)
--
-- = See Also
--
-- <https://www.khronos.org/registry/vulkan/specs/1.2-extensions/html/vkspec.html#VK_AMD_shader_core_properties2 VK_AMD_shader_core_properties2>,
-- 'ShaderCorePropertiesFlagsAMD',
-- 'Vulkan.Core10.Enums.StructureType.StructureType'
data PhysicalDeviceShaderCoreProperties2AMD = PhysicalDeviceShaderCoreProperties2AMD
{ -- | #features-shaderCoreFeatures# @shaderCoreFeatures@ is a bitmask of
-- 'ShaderCorePropertiesFlagBitsAMD' indicating the set of features
-- supported by the shader core.
shaderCoreFeatures :: ShaderCorePropertiesFlagsAMD
, -- | #limits-activeComputeUnitCount# @activeComputeUnitCount@ is an unsigned
-- integer value indicating the number of compute units that have been
-- enabled.
activeComputeUnitCount :: Word32
}
deriving (Typeable, Eq)
#if defined(GENERIC_INSTANCES)
deriving instance Generic (PhysicalDeviceShaderCoreProperties2AMD)
#endif
deriving instance Show PhysicalDeviceShaderCoreProperties2AMD
instance ToCStruct PhysicalDeviceShaderCoreProperties2AMD where
withCStruct x f = allocaBytes 24 $ \p -> pokeCStruct p x (f p)
pokeCStruct p PhysicalDeviceShaderCoreProperties2AMD{..} f = do
poke ((p `plusPtr` 0 :: Ptr StructureType)) (STRUCTURE_TYPE_PHYSICAL_DEVICE_SHADER_CORE_PROPERTIES_2_AMD)
poke ((p `plusPtr` 8 :: Ptr (Ptr ()))) (nullPtr)
poke ((p `plusPtr` 16 :: Ptr ShaderCorePropertiesFlagsAMD)) (shaderCoreFeatures)
poke ((p `plusPtr` 20 :: Ptr Word32)) (activeComputeUnitCount)
f
cStructSize = 24
cStructAlignment = 8
pokeZeroCStruct p f = do
poke ((p `plusPtr` 0 :: Ptr StructureType)) (STRUCTURE_TYPE_PHYSICAL_DEVICE_SHADER_CORE_PROPERTIES_2_AMD)
poke ((p `plusPtr` 8 :: Ptr (Ptr ()))) (nullPtr)
poke ((p `plusPtr` 16 :: Ptr ShaderCorePropertiesFlagsAMD)) (zero)
poke ((p `plusPtr` 20 :: Ptr Word32)) (zero)
f
instance FromCStruct PhysicalDeviceShaderCoreProperties2AMD where
peekCStruct p = do
shaderCoreFeatures <- peek @ShaderCorePropertiesFlagsAMD ((p `plusPtr` 16 :: Ptr ShaderCorePropertiesFlagsAMD))
activeComputeUnitCount <- peek @Word32 ((p `plusPtr` 20 :: Ptr Word32))
pure $ PhysicalDeviceShaderCoreProperties2AMD
shaderCoreFeatures activeComputeUnitCount
instance Storable PhysicalDeviceShaderCoreProperties2AMD where
sizeOf ~_ = 24
alignment ~_ = 8
peek = peekCStruct
poke ptr poked = pokeCStruct ptr poked (pure ())
instance Zero PhysicalDeviceShaderCoreProperties2AMD where
zero = PhysicalDeviceShaderCoreProperties2AMD
zero
zero
type ShaderCorePropertiesFlagsAMD = ShaderCorePropertiesFlagBitsAMD
-- | VkShaderCorePropertiesFlagBitsAMD - Bitmask specifying shader core
-- properties
--
-- = See Also
--
-- <https://www.khronos.org/registry/vulkan/specs/1.2-extensions/html/vkspec.html#VK_AMD_shader_core_properties2 VK_AMD_shader_core_properties2>,
-- 'PhysicalDeviceShaderCoreProperties2AMD', 'ShaderCorePropertiesFlagsAMD'
newtype ShaderCorePropertiesFlagBitsAMD = ShaderCorePropertiesFlagBitsAMD Flags
deriving newtype (Eq, Ord, Storable, Zero, Bits, FiniteBits)
conNameShaderCorePropertiesFlagBitsAMD :: String
conNameShaderCorePropertiesFlagBitsAMD = "ShaderCorePropertiesFlagBitsAMD"
enumPrefixShaderCorePropertiesFlagBitsAMD :: String
enumPrefixShaderCorePropertiesFlagBitsAMD = ""
showTableShaderCorePropertiesFlagBitsAMD :: [(ShaderCorePropertiesFlagBitsAMD, String)]
showTableShaderCorePropertiesFlagBitsAMD = []
instance Show ShaderCorePropertiesFlagBitsAMD where
showsPrec = enumShowsPrec enumPrefixShaderCorePropertiesFlagBitsAMD
showTableShaderCorePropertiesFlagBitsAMD
conNameShaderCorePropertiesFlagBitsAMD
(\(ShaderCorePropertiesFlagBitsAMD x) -> x)
(\x -> showString "0x" . showHex x)
instance Read ShaderCorePropertiesFlagBitsAMD where
readPrec = enumReadPrec enumPrefixShaderCorePropertiesFlagBitsAMD
showTableShaderCorePropertiesFlagBitsAMD
conNameShaderCorePropertiesFlagBitsAMD
ShaderCorePropertiesFlagBitsAMD
type AMD_SHADER_CORE_PROPERTIES_2_SPEC_VERSION = 1
-- No documentation found for TopLevel "VK_AMD_SHADER_CORE_PROPERTIES_2_SPEC_VERSION"
pattern AMD_SHADER_CORE_PROPERTIES_2_SPEC_VERSION :: forall a . Integral a => a
pattern AMD_SHADER_CORE_PROPERTIES_2_SPEC_VERSION = 1
type AMD_SHADER_CORE_PROPERTIES_2_EXTENSION_NAME = "VK_AMD_shader_core_properties2"
-- No documentation found for TopLevel "VK_AMD_SHADER_CORE_PROPERTIES_2_EXTENSION_NAME"
pattern AMD_SHADER_CORE_PROPERTIES_2_EXTENSION_NAME :: forall a . (Eq a, IsString a) => a
pattern AMD_SHADER_CORE_PROPERTIES_2_EXTENSION_NAME = "VK_AMD_shader_core_properties2"
| expipiplus1/vulkan | src/Vulkan/Extensions/VK_AMD_shader_core_properties2.hs | bsd-3-clause | 9,673 | 0 | 14 | 1,698 | 1,307 | 798 | 509 | -1 | -1 |
module Text.Highlighter.Lexer (runLexer) where
import Control.Monad.Except (ExceptT, runExceptT, throwError, catchError)
import Control.Monad.State (State, gets, modify, evalState)
import Text.Regex.PCRE.Light hiding (compile)
import Text.Regex.PCRE.Light.Char8 (compile)
import qualified Data.ByteString as BS
import Data.Sequence (Seq, empty, singleton, (><), viewl, null, ViewL(..))
import Data.Monoid ((<>))
import Control.Applicative ((<$>))
import Data.Foldable (toList, foldr1, mapM_)
import Prelude hiding (lex, foldr1, mapM_, concat, head, drop, tail, reverse, dropWhile, null)
import qualified Prelude as P
import Text.Highlighter.Types
data LexerState =
LexerState
{ lsLexer :: Lexer
, lsInput :: BS.ByteString
, lsState :: [TokenMatcher]
, lsLexed :: (Seq Token)
, lastNotNull :: Bool
}
deriving Show
type LexerM = ExceptT LexerError (State LexerState)
data LexerError
= NoMatchFor BS.ByteString
| OtherLexerError String
deriving Show
runLexer :: Lexer -> BS.ByteString -> Either LexerError [Token]
runLexer l s = toList <$> runLexer' l s
runLexer' :: Lexer -> BS.ByteString -> Either LexerError (Seq Token)
runLexer' l s = evalState (runExceptT lex) (LexerState l s [lStart l] empty True)
lex :: LexerM (Seq Token)
lex = do
done <- gets (BS.null . lsInput)
if done
then gets lsLexed
else do
ms <- getState
ts <- tryAll ms
if null ts || (BS.null . tText . head $ ts)
then modify $ \ls -> ls { lsLexed = lsLexed ls >< ts }
else modify $ \ls -> ls { lsLexed = lsLexed ls >< ts
, lastNotNull = (BS.last . tText . head $ ts) == 10
}
lex
where
getState = gets (P.head . lsState)
isBOL :: LexerM Bool
isBOL = gets lastNotNull
head :: Seq a -> a
head x = let (b :< _) = viewl x
in b
tryAll :: [Match] -> LexerM (Seq Token)
tryAll [] = do
i <- gets lsInput
throwError (NoMatchFor i)
tryAll (AnyOf ms:ms') =
tryAll (ms ++ ms')
tryAll (m:ms) = do
atbol <- isBOL
fs <- gets (lFlags . lsLexer)
let opts
| atbol = [exec_anchored]
| otherwise = [exec_anchored, exec_notbol]
i <- gets lsInput
case match (compile (mRegexp m) fs) i opts of
Just [] -> do
nextState (mNextState m) []
return empty
Just (s:ss) -> do
modify $ \ls -> ls { lsInput = BS.drop (BS.length s) i }
nextState (mNextState m) (s:ss)
toTokens (s:ss) (mType m)
Nothing ->
tryAll ms `catchError` trySkipping
where
trySkipping (NoMatchFor _) = tryAllFirst (m:ms)
trySkipping e = throwError e
tryAllFirst :: [Match] -> LexerM (Seq Token)
tryAllFirst [] = do
i <- gets lsInput
throwError (NoMatchFor i)
tryAllFirst (AnyOf ms:ms') =
tryAllFirst (ms ++ ms')
tryAllFirst (m:ms) = do
atbol <- isBOL
fs <- gets (lFlags . lsLexer)
let opts
| atbol = []
| otherwise = [exec_notbol]
i <- gets lsInput
case match (compile (mRegexp m) fs) i opts of
Just (s:ss) -> do
let (skipped, next) = skipFailed i s
modify $ \ls -> ls { lsInput = next }
ts <- toTokens (s:ss) (mType m)
return . singleton . Token Error $ (skipped <> (tText $ head ts))
_ -> tryAllFirst ms
toTokens :: [BS.ByteString] -> TokenType -> LexerM (Seq Token)
toTokens (s:_) (Using l) = either throwError return (runLexer' l s)
toTokens (_:ss) (ByGroups ts) = foldr1 (><) <$> mapM (\(s,t) -> toTokens [s] t) (P.zip ss ts)
toTokens (s:_) t = return $ singleton $ Token t s
toTokens [] _ = return empty
-- Given the starting point, return the text preceding and after
-- the failing regexp match
skipFailed :: BS.ByteString -> BS.ByteString -> (BS.ByteString, BS.ByteString)
skipFailed i r
| r `BS.isPrefixOf` i = (BS.empty, BS.drop (BS.length r) i)
| otherwise =
let (pre, next) = skipFailed (BS.tail i) r
in (BS.cons (BS.head i) pre, next)
nextState :: NextState -> [BS.ByteString] -> LexerM ()
nextState Continue _ = return ()
nextState Pop _ =
modify $ \ls -> ls { lsState = P.tail (lsState ls) }
nextState (PopNum n) _ =
modify $ \ls -> ls { lsState = P.drop n (lsState ls) }
nextState Push _ =
modify $ \ls -> ls { lsState = P.head (lsState ls) : lsState ls }
nextState (GoTo n) _ =
modify $ \ls -> ls { lsState = n : lsState ls }
nextState (CapturesTo f) cs =
modify $ \ls -> ls { lsState = f (map fromBS cs) : lsState ls }
where
fromBS = map (toEnum . fromEnum) . BS.unpack
nextState (DoAll nss) cs = mapM_ (flip nextState cs) nss
nextState (Combined nss) _ =
modify $ \ls -> ls { lsState = P.concat nss : lsState ls }
| chemist/highlighter | src/Text/Highlighter/Lexer.hs | bsd-3-clause | 4,808 | 0 | 20 | 1,313 | 2,018 | 1,050 | 968 | -1 | -1 |
module Problem15 where
--
-- Problem 15: Lattice paths
--
-- Starting in the top left corner of a 2×2 grid, and only being able to move to
-- the right and down, there are exactly 6 routes to the bottom right corner.
--
-- https://projecteuler.net/project/images/p015.gif
--
-- Paths: (R=Right, D=Down)
--
-- R R D D
-- R D R D
-- R D D R
-- D R R D
-- D R D R
-- D D R R
--
-- |R| = |D|
--
-- How many such routes are there through a 20×20 grid?
--
-- 40! / 20!20!
| c0deaddict/project-euler | src/Part1/Problem15.hs | bsd-3-clause | 469 | 0 | 2 | 105 | 26 | 25 | 1 | 1 | 0 |
{-# LANGUAGE MultiParamTypeClasses #-}
-- |
-- Module : Simulation.Aivika.Experiment.Base.InfoView
-- Copyright : Copyright (c) 2012-2017, David Sorokin <[email protected]>
-- License : BSD3
-- Maintainer : David Sorokin <[email protected]>
-- Stability : experimental
-- Tested with: GHC 8.0.1
--
-- The module defines 'InfoView' that shows the description of series.
--
module Simulation.Aivika.Experiment.Base.InfoView
(InfoView(..),
defaultInfoView) where
import Control.Monad
import Control.Monad.Trans
import Control.Concurrent.MVar
import Data.IORef
import Data.Maybe
import Data.Monoid
import Simulation.Aivika
import Simulation.Aivika.Experiment.Types
import Simulation.Aivika.Experiment.Base.WebPageRenderer
import Simulation.Aivika.Experiment.Base.ExperimentWriter
import Simulation.Aivika.Experiment.Base.HtmlWriter
import Simulation.Aivika.Experiment.Concurrent.MVar
-- | Defines the 'View' that shows the description of series.
data InfoView =
InfoView { infoTitle :: String,
-- ^ This is a title for the view.
infoDescription :: String,
-- ^ This is a text description used in HTML.
infoTransform :: ResultTransform,
-- ^ The transform applied to the results before receiving series.
infoSeries :: ResultTransform
-- ^ It defines the series for which the description is shown.
}
-- | The default description view.
defaultInfoView :: InfoView
defaultInfoView =
InfoView { infoTitle = "Information",
infoDescription = "It shows the information about simulation entities:",
infoTransform = id,
infoSeries = id }
instance ExperimentView InfoView (WebPageRenderer a) where
outputView v =
let reporter exp renderer dir =
do st <- newInfo v exp dir
let context =
WebPageContext $
WebPageWriter { reporterWriteTOCHtml = infoTOCHtml st,
reporterWriteHtml = infoHtml st }
return ExperimentReporter { reporterInitialise = return (),
reporterFinalise = return (),
reporterSimulate = simulateInfo st,
reporterContext = context }
in ExperimentGenerator { generateReporter = reporter }
-- | The state of the view.
data InfoViewState =
InfoViewState { infoView :: InfoView,
infoExperiment :: Experiment,
infoResults :: MVar (Maybe InfoResults) }
-- | The information table.
data InfoResults =
InfoResults { infoNames :: [String],
infoValues :: [String] }
-- | Create a new state of the view.
newInfo :: InfoView -> Experiment -> FilePath -> ExperimentWriter InfoViewState
newInfo view exp dir =
do r <- liftIO $ newMVar Nothing
return InfoViewState { infoView = view,
infoExperiment = exp,
infoResults = r }
-- | Create a new information table.
newInfoResults :: [ResultSource] -> ResultLocalisation -> Experiment -> IO InfoResults
newInfoResults sources loc exp =
do let xs =
flip map sources $ \source ->
case source of
ResultItemSource (ResultItem x) ->
[(resultNameToTitle $ resultItemName x,
localiseResultDescription loc $ resultItemId x)]
ResultObjectSource x ->
[(resultNameToTitle $ resultObjectName x,
localiseResultDescription loc $ resultObjectId x)]
ResultVectorSource x ->
[(resultNameToTitle $ resultVectorName x,
localiseResultDescription loc $ resultVectorId x)]
ResultSeparatorSource x ->
[]
(names, values) = unzip $ concat xs
return InfoResults { infoNames = names,
infoValues = values }
-- | Require to return the unique information table associated with the specified state.
requireInfoResults :: InfoViewState -> [ResultSource] -> IO InfoResults
requireInfoResults st sources =
let view = infoView st
exp = infoExperiment st
loc = experimentLocalisation exp
in maybePutMVar (infoResults st)
(newInfoResults sources loc exp) $ \results ->
do let xs =
flip map sources $ \source ->
case source of
ResultItemSource (ResultItem x) ->
[resultNameToTitle $ resultItemName x]
ResultObjectSource x ->
[resultNameToTitle $ resultObjectName x]
ResultVectorSource x ->
[resultNameToTitle $ resultVectorName x]
ResultSeparatorSource x ->
[]
let names = concat xs
if (names /= infoNames results)
then error "Series with different names are returned for different runs: requireInfoResults"
else return results
-- | Simulate the specified series.
simulateInfo :: InfoViewState -> ExperimentData -> Composite ()
simulateInfo st expdata =
do let view = infoView st
rs = infoSeries view $
infoTransform view $
experimentResults expdata
sources = resultSourceList rs
liftIO $ requireInfoResults st sources
return ()
-- | Get the HTML code.
infoHtml :: InfoViewState -> Int -> HtmlWriter ()
infoHtml st index =
do header st index
results <- liftIO $ readMVar (infoResults st)
case results of
Nothing -> return ()
Just results ->
do let names = infoNames results
values = infoValues results
writeHtmlList $
forM_ (zip names values) $ \(name, value) ->
writeHtmlListItem $
do writeHtmlText name
writeHtmlText " - "
writeHtmlText value
header :: InfoViewState -> Int -> HtmlWriter ()
header st index =
do writeHtmlHeader3WithId ("id" ++ show index) $
writeHtmlText (infoTitle $ infoView st)
let description = infoDescription $ infoView st
unless (null description) $
writeHtmlParagraph $
writeHtmlText description
-- | Get the TOC item.
infoTOCHtml :: InfoViewState -> Int -> HtmlWriter ()
infoTOCHtml st index =
writeHtmlListItem $
writeHtmlLink ("#id" ++ show index) $
writeHtmlText (infoTitle $ infoView st)
| dsorokin/aivika-experiment | Simulation/Aivika/Experiment/Base/InfoView.hs | bsd-3-clause | 6,535 | 0 | 21 | 2,037 | 1,330 | 693 | 637 | 131 | 5 |
{-# LANGUAGE FlexibleContexts #-}
{-# LANGUAGE OverloadedStrings #-}
module Jade.TopLevel ( getNetsWithName
, connectWiresWithSameSigName
, dependencyOrder
, explodeConnect
, getAllIndexesWithName
, getAllSigNames
, getInputs
, getInternalSigNames
, getModule
, getOutputs
, getSchematic
, getSubModules
, getTerminalsAtPoint
, getWidthOfValName
, getWidthOfPartsAtTerminal
, isNetDriver
, netWithTerminal
, nets
, numNets
, replicationDepth
, terminals
) where
import Control.Monad
import Control.Monad.State
import qualified Data.List as DL
import qualified Data.Map as DM
import Data.Maybe
import qualified Data.Maybe as Maybe
import qualified Data.Set as DS
import qualified Data.Vector as DV
import qualified Jade.Decode.Decode as D
import qualified Jade.Jumper as Jumper
import qualified Jade.MemUnit as MemUnit
import qualified Jade.Module as Module
import qualified Jade.Net as Net
import qualified Jade.Part as Part
import qualified Jade.Term as Term
import qualified Jade.Schematic as Schem
import Jade.Common
import qualified Jade.UnionFindST as UF
import qualified Jade.Wire as Wire
import qualified Web.Hashids as WH
import qualified Jade.Decode.Bundle as Bundle
getSubModules :: String -> J [SubModule]
getSubModules modname = do --"TopLevel.getSubModules" <? do
(Module _ schem _ _) <- getModule modname
case schem of
Just schem -> return $ Schem.getSubModules schem
Nothing -> die "No schematics found"
-- a function to possible create an edge given a wire and a part
makePartEdge :: Wire -> Part -> J (Maybe Edge)
makePartEdge wire part = "TopLevel.makePartEdge" <? do
let (loc1, loc2) = Wire.ends wire
ploc <- Part.loc part
if ploc == loc1
then return $ Just $ Edge (Node loc1 (WireC wire)) (Node ploc part)
else if ploc == loc2
then return $ Just $ Edge (Node loc2 (WireC wire)) (Node ploc part)
else return Nothing
makeWire2WireEdge :: Wire -> Wire -> J (Maybe Edge)
makeWire2WireEdge w1 w2 = "TopLevel.makeWire2WireEdge" <?
let (loc1, loc2) = Wire.ends w1
(loc3, loc4) = Wire.ends w2
econ p1 v p2 w = Just $ Edge (Node p1 (WireC v)) (Node p2 (WireC w))
in return $ case (loc1 == loc3, loc1 == loc4, loc2 == loc3, loc2 == loc4) of
(True, _, _, _) -> econ loc1 w1 loc3 w2
(_, True, _, _) -> econ loc1 w1 loc4 w2
(_, _, True, _) -> econ loc2 w1 loc3 w2
(_, _, _, True) -> econ loc2 w1 loc4 w2
_ -> Nothing
terminalsOverlapP (Terminal (Coord3 x1 y1 _) _) (Terminal (Coord3 x2 y2 _) _)
= (x1,y1) == (x2,y2)
getOverlappingTerminals :: String -> J [(Terminal, Terminal)]
getOverlappingTerminals modname = "TopLevel.collectOverlappingTerminals" <? do
allsubs <- getSubModules modname
ts <- concatMapM terminals allsubs
return $ DL.nub $ DL.sort [ (min t1 t2, max t1 t2) | t1 <- ts, t2 <- ts,
terminalsOverlapP t1 t2,
t1 /= t2 ]
connectOverlappingTerminals :: [(Terminal, Terminal)] -> J [Wire]
connectOverlappingTerminals termPairs = "TopLevel.connectOverlappingTerminals" <? do
return $ [Wire.mkDegenerate c | (Terminal c _, _) <- termPairs]
processEdges :: [Wire] -> [Part] -> J [Edge]
processEdges wires parts = "TopLevel.processEdges" <? do
nb "with all wires, make an edge from ones that share a point with a part"
partEdges <- sequence [makePartEdge w p | w <- wires, p <- parts]
let Just edges = sequence $ filter Maybe.isJust partEdges
wireNbrs <- sequence [makeWire2WireEdge v w | (v, w) <- triangleProd wires]
let Just nbrs = sequence $ filter Maybe.isJust wireNbrs
return $ edges ++ nbrs
makeJumperWire :: Jumper -> J Wire
makeJumperWire jumper = "TopLevel.makeJumperEdge" <? do
nb "find the endpoints of the jumper"
let (p1, p2) = Jumper.getEnds jumper
nb "create a wire where the jumper is"
return $ Wire.new p1 p2
makePortWire :: Port -> J Wire
makePortWire (Port (Coord3 x y r) sig) = "TopLevel.makePortWire" <? do
nb "create a wire of length zero, that has the signal from the port"
return $ Wire (Coord5 x y r 0 0) sig
connectSubWires :: Wire -> Wire -> J (Maybe Wire)
connectSubWires w1 w2 = "TopLevel.connectSubWires" <? do
assert (Wire.width w1 == 1) "this function assumes wires are width 1"
assert (Wire.width w2 == 1) "this function assumes wires are width 1"
assert (Wire.hasSigName w1) "can't be literal"
assert (Wire.hasSigName w2) "can't be literal"
if w1 `Wire.hasSameSig` w2
then do nb "--------------------------------------------"
enb w1
enb w2
return $ Just $ Wire.new (fst $ Wire.ends w1) (fst $ Wire.ends w2)
else return Nothing
explodeConnect :: (Wire, Wire) -> J [Wire]
explodeConnect (w1, w2) = "TopLevel.explodeConnect" <? do
let subwires1 = Wire.explode w1
subwires2 = Wire.explode w2
enb (subwires1)
enb (subwires2)
catMaybes <$> sequence [connectSubWires sw1 sw2 | sw1 <- subwires1, sw2 <- subwires2 ]
connectWiresWithSameSigName :: [Part] -> J [Wire]
connectWiresWithSameSigName parts = "TopLevel.connectWiresWithSameSigName" <? do
let wires = triangleProd $ filter Wire.hasSigName $ catMaybes $ map Part.toWire parts
let pairs = [(w1, w2) | (w1, w2) <- wires, w1 `Wire.hasSameSig` w2]
return [Wire.new (fst $ Wire.ends w1) (fst $ Wire.ends w2) | (w1, w2) <- pairs]
-- concat <$> sequence [explodeConnect (w1, w2) | (w1, w2) <- wires] --, w1 `Wire.hasSameSig` w2]
-- | What's going on here? Now that the decoder explodes signal names
-- out into val bundles immediately, it's possible to connect wire
-- names at the very beginning, rather it's possible to associate two
-- wires for the unionfind algorithm. this is the right way to do it.
-- The wrong way, is what I was doing, some after the fact kludge
-- braindead patching. So two parts enter this function. parts have
-- wires. wires have bundles. The change in procedure here, what the
-- whole refactoring was about, is to intersect these bundles and feed
-- them to union find. This will reduce the total number of nets!
-- which is good. wires should just magically connect if this works.
-- So, back to how union find distinguishes nodes, by
nets :: String -> J [Net]
nets modname = "TopLevel.nets" <? do
-- memoize, TODO: abstract this away.
Memo table <- getMemo
case DM.lookup modname table of
-- Already computed this net, so return it.
Just nets -> return nets
-- Compute the net, insert it into the memo map, then return the net
Nothing -> do cs <- nets' modname
putMemo $ Memo (DM.insert modname cs table)
return cs
nets' :: String -> J [Net]
nets' modname = do --"TopLevel.nets_" <? do
edges <- getEdges modname
let nets_ = UF.components $ edges
nb "let nets = UF.components $ edges ++ wireEdges"
enb nets_
return nets_
getEdges :: String -> J [Edge]
getEdges modname = "TopLevel.getEdges" <? do
nb "---------------------------------"
nbf "get the module: {0}" [modname]
(Module _ (Just schem@(Schematic parts)) _ _) <- getModule modname
terms <- sequence [terminals submod | submod <- Schem.getSubModules schem]
let wires = [w | WireC w <- parts]
ports = [p | PortC p <- parts]
jumpers = Schem.getJumpers schem
termcs = map TermC $ concat terms
ssnw <- connectWiresWithSameSigName parts
jumperWires <- mapM makeJumperWire jumpers
portWires <- mapM makePortWire ports
ts <- getOverlappingTerminals modname
overlappingTermWires <- connectOverlappingTerminals ts
let allWires = concat [ wires
, jumperWires
, ssnw
, portWires
, overlappingTermWires]
wireEdges = map Wire.toEdge allWires
edges <- processEdges allWires termcs
return (wireEdges ++ edges)
-- | VHDL requires that modules be instantiated in dependency order,
dependencyOrder :: String -> J [String]
dependencyOrder modname = "TopLevel.dependencyOrder" <?
if not $ modname `startsWith` "/user/" then return []
else do m <- getModule modname
schem <- Module.getSchematic m
let subnames = DL.nub [subname | (SubModule subname _) <- Schem.getSubModules schem]
children <- concatMapM dependencyOrder subnames
return $ filter (`startsWith` "/user") $ DL.nub $ children ++ subnames
-- |Get the graph net which contains the terminals.
netWithTerminal :: [Char] -> Terminal -> J Net
netWithTerminal modname term@(Terminal c3@(Coord3 x y _) _) = "TopLevel.netWithTerminal" <? do
nets <- nets modname
let result = filter (flip Net.hasTerm term) nets
case length result of
0 -> die $ concat [ " No net found in module: ", modname
, " that has a terminal: ", show term ]
1 -> return $ head result
x -> die $ concat [ show x, " nets found in module: ", modname
, " that has a terminal: ", show term, "."
, " This should not be possible, because all such nets should"
, " be connected if they contain the same node" ]
-- | Get a list of input and output terminals in a submodule offset by
-- the position of the submodule
terminals :: SubModule -> J [Terminal]
terminals (SubModule modname offset) = do --"TopLevel.terminals" <? do
nb $ show ("TopLevel.terminals checks submodule: " ++ modname)
mod <- getModule modname
Module.terminals mod offset
terminals (SubMemUnit memunit) = "TopLevel.terminals/memunit" <? do
MemUnit.terminals memunit
-- | Get the number of distinct nodes in the schematic
numNets :: String -> J Int
numNets modname = "TopLevel.numNets" <? do
length <$> nets modname ? "Couldn't get number of nets"
-- -- | Get the input of a module. This requires tests to be defined in
-- -- the module referenced, because .input directive of the test script
-- -- indicate the target signals in the schematic
getInputs :: String -> J Inputs
getInputs modname = "TopLevel.getInputs" <? do
getModule modname >>= Module.getInputs
-- -- | Get the outputs of a module. This requires tests to be defined in
-- -- the module referenced, because the .output directive of the test
-- -- script indicate the source signals in the schematic
getOutputs :: String -> J Outputs
getOutputs modname = "TopLevel.getOutputs" <? do
mod <- getModule modname
let msg = "TopLevel.getOutputs couldn't find outputs in module: " ++ modname
Module.getOutputs mod ? msg
-- | The assumption always is, that the jade module works and is
-- tested. With that in mind, then it's safe to assume that there is
-- one driving signal per net. This
-- function finds the driving signal for a given net.
-- If a signal is in more than one net then it is a driving signal.
-- | What signals are driving? .input signals from the test script
-- indicate a set of driving signals. OUTPUT terminals of sub modules
-- are also driving signals.
getNetWithTerminal :: String -> Terminal -> J Net
getNetWithTerminal modname term = "getNetWithTerminal" <? do
allNets <- nets modname
let matches = [c | c <- allNets, Net.hasTerm c term]
case matches of
[] -> die $ "No net found with terminal: " ++ show term
[c] -> do nb $ "found net with terminal: " ++ show term
return c
_ -> impossible $ "More than one net found with terminal: " ++ show term
getPartsConnectedToTerminal :: String -> Terminal -> J [Part]
getPartsConnectedToTerminal modname terminal = "TopLevel.getPartsConnectedToTerminal" <? do
let (Terminal (Coord3 x y _) _) = terminal
schem <- getSchematic modname
Schem.getAllPartsAtPoint schem (Point x y)
getWidthOfPartsAtTerminal :: String -> Terminal -> J Int
getWidthOfPartsAtTerminal modname terminal = "TopLevel.getWidthOfPartsAtTerminal" <? do
parts <- getPartsConnectedToTerminal modname terminal
maximum <$> mapM Part.width (Part.removeTerms parts)
getTerminalsAtPoint :: String -> Point -> J [Terminal]
getTerminalsAtPoint modname point@(Point x1 y1) = "TopLevel.getTerminalAtPoint" <? do
schem <- getSchematic modname
let subs = Schem.getSubModules schem
allTerms <- concatMapM terminals subs
return $ filter (flip Term.atPoint point) allTerms
getRatio :: Terminal -> Part -> J (Int, Int)
getRatio terminal part = "TopLevel.getTerminalRatio" <? do
let tw = Term.width terminal
pw <- Part.width part
return (tw, pw)
getRatios modname term = do
parts <- getPartsConnectedToTerminal modname term
mapM (getRatio term) (filter (not . Part.isSubModule) parts)
replicationDepth :: String -> SubModule -> J Int
replicationDepth modname submod = "TopLevel.replicationDepth" <? do
terms <- terminals submod
ratios <- concatMapM (getRatios modname) terms
let numReps = [if tw >= pw then 1 else pw `div` tw | (tw, pw) <- ratios]
return $ maximum numReps
getNetsWithName :: String -> String -> J [Net]
getNetsWithName modname signame = "TopLevel.getNetsWithName" <? do
ns <- nets modname
filterM (flip Net.containsIdent signame) ns
getAllSigNames :: String -> J [String]
getAllSigNames modname = do
nets <- nets modname
let parts = (concat $ map Net.parts nets)
results = concat $ map Part.getNames parts
return $ DL.nub results
-- | Given a valnal name scour all the nets that contains the
-- name for the total width of all valnals with the name.
getWidthOfValName :: String -> String -> J Int
getWidthOfValName modname valname = "TopLevel.getWidthOfValName" <? do
nets <- getNetsWithName modname valname
vals <- concatMapM (flip Net.getValsWithIdent valname) nets
return $ length (DL.nub vals)
getInternalSigNames modname = "TopLevel.getInternalSigNames" <? do
m <- getModule modname
allNames <- getAllSigNames modname
(Inputs inputBundles) <- getInputs modname
(Outputs outputBundles) <- getOutputs modname
let inputNames = concat $ map Bundle.getNames inputBundles
outputNames = concat $ map Bundle.getNames outputBundles
return $ allNames DL.\\ (inputNames ++ outputNames)
getAllIndexesWithName :: String -> String -> J [Val]
getAllIndexesWithName modname name = "TopLevel.getAllIndexesWithName" <? do
allNets <- nets modname
return $ DL.nub $ concat $ map (flip Net.getIndexesWithName name) allNets
isNetDriver :: String -> Net -> J Bool
isNetDriver modname net = "TopLevel.isNetDriven" <? do
-- does this net contain names that are included in the .input
-- groups? does this net contain nodes that are connected to the
-- output of contradiction. if an internal name is both on the
-- output and input of a submodule, then unfortunately this means
-- that the net actually belongs to both. but wait. net analysis is pass#1 analysis
-- so, no, they actually aren't.
-- how many nets are there here?
return False
| drhodes/jade2hdl | src/Jade/TopLevel.hs | bsd-3-clause | 15,284 | 0 | 17 | 3,628 | 3,972 | 1,996 | 1,976 | 268 | 5 |
{-# LANGUAGE FlexibleInstances #-}
{-# LANGUAGE OverloadedStrings #-}
{-# LANGUAGE RecordWildCards #-}
{-# LANGUAGE TemplateHaskell #-}
module Main where
------------------------------------------------------------------------------
import Control.Monad.Trans
import Data.ByteString (ByteString)
import Control.Lens
import qualified Data.Text.Encoding as T
import qualified Database.Persist as P
import Database.Persist.Sql
import Snap
import Snap.Snaplet.Auth
import Snap.Snaplet.Auth.Backends.Persistent
import Snap.Snaplet.Persistent
import Snap.Snaplet.Session
import Snap.Snaplet.Session.Backends.CookieSession
------------------------------------------------------------------------------
data App = App
{ _sess :: Snaplet SessionManager
, _db :: Snaplet PersistState
, _auth :: Snaplet (AuthManager App)
}
makeLenses ''App
instance HasPersistPool (Handler b App) where
getPersistPool = with db getPersistPool
------------------------------------------------------------------------------
-- | The application's routes.
routes :: [(ByteString, Handler App App ())]
routes = [ ("/", writeText "hello")
, ("foo", fooHandler)
, ("add/:uname", addHandler)
]
fooHandler :: Handler App App ()
fooHandler = do
results <- runPersist $ P.selectList [] []
liftIO $ print (map db2au results)
addHandler :: Handler App App ()
addHandler = do
mname <- getParam "uname"
let name = maybe "guest" T.decodeUtf8 mname
u <- with auth $ createUser name ""
liftIO $ print u
------------------------------------------------------------------------------
-- | The application initializer.
app :: SnapletInit App App
app = makeSnaplet "app" "An snaplet example application." Nothing $ do
s <- nestSnaplet "" sess $
initCookieSessionManager "site_key.txt" "_cookie" Nothing Nothing
d <- nestSnaplet "db" db $ initPersist (runMigrationUnsafe migrateAuth)
a <- nestSnaplet "auth" auth $
initPersistAuthManager sess (persistPool $ view snapletValue d)
addRoutes routes
return $ App s d a
main :: IO ()
main = serveSnaplet defaultConfig app
| Soostone/snaplet-persistent | example/Site.hs | bsd-3-clause | 2,261 | 0 | 13 | 481 | 513 | 273 | 240 | 49 | 1 |
{-# LANGUAGE PatternGuards #-}
-----------------------------------------------------------------------------
-- |
-- Module : Distribution.Lex
-- Copyright : Ben Gamari 2015-2019
--
-- Maintainer : [email protected]
-- Portability : portable
--
-- This module contains a simple lexer supporting quoted strings
module Distribution.Lex (
tokenizeQuotedWords
) where
import Prelude ()
import Distribution.Compat.Prelude
newtype DList a = DList ([a] -> [a])
runDList :: DList a -> [a]
runDList (DList run) = run []
singleton :: a -> DList a
singleton a = DList (a:)
instance Monoid (DList a) where
mempty = DList id
mappend = (<>)
instance Semigroup (DList a) where
DList a <> DList b = DList (a . b)
tokenizeQuotedWords :: String -> [String]
tokenizeQuotedWords = filter (not . null) . go False mempty
where
go :: Bool -- ^ in quoted region
-> DList Char -- ^ accumulator
-> String -- ^ string to be parsed
-> [String] -- ^ parse result
go _ accum []
| [] <- accum' = []
| otherwise = [accum']
where accum' = runDList accum
go False accum (c:cs)
| isSpace c = runDList accum : go False mempty cs
| c == '"' = go True accum cs
go True accum (c:cs)
| c == '"' = go False accum cs
go quoted accum (c:cs)
= go quoted (accum `mappend` singleton c) cs
| sopvop/cabal | Cabal/Distribution/Lex.hs | bsd-3-clause | 1,409 | 0 | 11 | 378 | 437 | 230 | 207 | 32 | 4 |
{-# LANGUAGE GADTs #-}
{-# LANGUAGE RankNTypes #-}
{-# LANGUAGE StandaloneDeriving #-}
data StackItem a where
Snum :: forall a. Fractional a => a -> StackItem a
Sop :: OpDesc -> StackItem a
deriving instance Show a => Show (StackItem a)
-- AZ added to test Trac #10399
data MaybeDefault v where
SetTo :: forall v . ( Eq v, Show v ) => !v -> MaybeDefault v
SetTo4 :: forall v a. (( Eq v, Show v ) => v -> MaybeDefault v -> a -> MaybeDefault [a])
| mpickering/ghc-exactprint | tests/examples/ghc710/GADTContext.hs | bsd-3-clause | 481 | 9 | 12 | 125 | 157 | 85 | 72 | 12 | 0 |
module Matterhorn.Events.ShowHelp where
import Prelude ()
import Matterhorn.Prelude
import Brick
import qualified Graphics.Vty as Vty
import Matterhorn.Constants
import Matterhorn.Events.Keybindings
import Matterhorn.Types
onEventShowHelp :: Vty.Event -> MH Bool
onEventShowHelp =
handleKeyboardEvent helpKeybindings $ \ e -> case e of
Vty.EvKey _ _ -> popMode
_ -> return ()
helpKeybindings :: KeyConfig -> KeyHandlerMap
helpKeybindings = mkKeybindings helpKeyHandlers
helpKeyHandlers :: [KeyEventHandler]
helpKeyHandlers =
[ mkKb ScrollUpEvent "Scroll up" $
mh $ vScrollBy (viewportScroll HelpViewport) (-1)
, mkKb ScrollDownEvent "Scroll down" $
mh $ vScrollBy (viewportScroll HelpViewport) 1
, mkKb PageUpEvent "Page up" $
mh $ vScrollBy (viewportScroll HelpViewport) (-1 * pageAmount)
, mkKb PageDownEvent "Page down" $
mh $ vScrollBy (viewportScroll HelpViewport) (1 * pageAmount)
, mkKb CancelEvent "Return to the previous interface" $
popMode
, mkKb ScrollBottomEvent "Scroll to the end of the help" $
mh $ vScrollToEnd (viewportScroll HelpViewport)
, mkKb ScrollTopEvent "Scroll to the beginning of the help" $
mh $ vScrollToBeginning (viewportScroll HelpViewport)
]
popMode :: MH ()
popMode = do
ShowHelp _ prevMode <- use (csCurrentTeam.tsMode)
setMode prevMode
| matterhorn-chat/matterhorn | src/Matterhorn/Events/ShowHelp.hs | bsd-3-clause | 1,447 | 0 | 11 | 345 | 366 | 189 | 177 | 35 | 2 |
-----------------------------------------------------------------------------
-- |
-- Module : Distribution.Simple.SrcDist
-- Copyright : Simon Marlow 2004
-- License : BSD3
--
-- Maintainer : [email protected]
-- Portability : portable
--
-- This handles the @sdist@ command. The module exports an 'sdist' action but
-- also some of the phases that make it up so that other tools can use just the
-- bits they need. In particular the preparation of the tree of files to go
-- into the source tarball is separated from actually building the source
-- tarball.
--
-- The 'createArchive' action uses the external @tar@ program and assumes that
-- it accepts the @-z@ flag. Neither of these assumptions are valid on Windows.
-- The 'sdist' action now also does some distribution QA checks.
-- NOTE: FIX: we don't have a great way of testing this module, since
-- we can't easily look inside a tarball once its created.
module Distribution.Simple.SrcDist (
-- * The top level action
sdist,
-- ** Parts of 'sdist'
printPackageProblems,
prepareTree,
createArchive,
-- ** Snapshots
prepareSnapshotTree,
snapshotPackage,
snapshotVersion,
dateToSnapshotNumber,
-- * Extracting the source files
listPackageSources
) where
import Prelude ()
import Distribution.Compat.Prelude
import Distribution.PackageDescription hiding (Flag)
import Distribution.PackageDescription.Check hiding (doesFileExist)
import Distribution.Package
import Distribution.ModuleName
import qualified Distribution.ModuleName as ModuleName
import Distribution.Version
import Distribution.Simple.Utils
import Distribution.Simple.Setup
import Distribution.Simple.PreProcess
import Distribution.Simple.LocalBuildInfo
import Distribution.Simple.BuildPaths
import Distribution.Simple.Program
import Distribution.Text
import Distribution.Verbosity
import Data.List (partition)
import qualified Data.Map as Map
import Data.Time (UTCTime, getCurrentTime, toGregorian, utctDay)
import System.Directory ( doesFileExist )
import System.IO (IOMode(WriteMode), hPutStrLn, withFile)
import System.FilePath ((</>), (<.>), dropExtension, isRelative)
-- |Create a source distribution.
sdist :: PackageDescription -- ^information from the tarball
-> Maybe LocalBuildInfo -- ^Information from configure
-> SDistFlags -- ^verbosity & snapshot
-> (FilePath -> FilePath) -- ^build prefix (temp dir)
-> [PPSuffixHandler] -- ^ extra preprocessors (includes suffixes)
-> IO ()
sdist pkg mb_lbi flags mkTmpDir pps =
-- When given --list-sources, just output the list of sources to a file.
case (sDistListSources flags) of
Flag path -> withFile path WriteMode $ \outHandle -> do
(ordinary, maybeExecutable) <- listPackageSources verbosity pkg pps
traverse_ (hPutStrLn outHandle) ordinary
traverse_ (hPutStrLn outHandle) maybeExecutable
notice verbosity $ "List of package sources written to file '"
++ path ++ "'"
NoFlag -> do
-- do some QA
printPackageProblems verbosity pkg
when (isNothing mb_lbi) $
warn verbosity "Cannot run preprocessors. Run 'configure' command first."
date <- getCurrentTime
let pkg' | snapshot = snapshotPackage date pkg
| otherwise = pkg
case flagToMaybe (sDistDirectory flags) of
Just targetDir -> do
generateSourceDir targetDir pkg'
info verbosity $ "Source directory created: " ++ targetDir
Nothing -> do
createDirectoryIfMissingVerbose verbosity True tmpTargetDir
withTempDirectory verbosity tmpTargetDir "sdist." $ \tmpDir -> do
let targetDir = tmpDir </> tarBallName pkg'
generateSourceDir targetDir pkg'
targzFile <- createArchive verbosity pkg' mb_lbi tmpDir targetPref
notice verbosity $ "Source tarball created: " ++ targzFile
where
generateSourceDir targetDir pkg' = do
setupMessage verbosity "Building source dist for" (packageId pkg')
prepareTree verbosity pkg' mb_lbi targetDir pps
when snapshot $
overwriteSnapshotPackageDesc verbosity pkg' targetDir
verbosity = fromFlag (sDistVerbosity flags)
snapshot = fromFlag (sDistSnapshot flags)
distPref = fromFlag $ sDistDistPref flags
targetPref = distPref
tmpTargetDir = mkTmpDir distPref
-- | List all source files of a package. Returns a tuple of lists: first
-- component is a list of ordinary files, second one is a list of those files
-- that may be executable.
listPackageSources :: Verbosity -- ^ verbosity
-> PackageDescription -- ^ info from the cabal file
-> [PPSuffixHandler] -- ^ extra preprocessors (include
-- suffixes)
-> IO ([FilePath], [FilePath])
listPackageSources verbosity pkg_descr0 pps = do
-- Call helpers that actually do all work.
ordinary <- listPackageSourcesOrdinary verbosity pkg_descr pps
maybeExecutable <- listPackageSourcesMaybeExecutable pkg_descr
return (ordinary, maybeExecutable)
where
pkg_descr = filterAutogenModules pkg_descr0
-- | List those source files that may be executable (e.g. the configure script).
listPackageSourcesMaybeExecutable :: PackageDescription -> IO [FilePath]
listPackageSourcesMaybeExecutable pkg_descr =
-- Extra source files.
fmap concat . for (extraSrcFiles pkg_descr) $ \fpath -> matchFileGlob fpath
-- | List those source files that should be copied with ordinary permissions.
listPackageSourcesOrdinary :: Verbosity
-> PackageDescription
-> [PPSuffixHandler]
-> IO [FilePath]
listPackageSourcesOrdinary verbosity pkg_descr pps =
fmap concat . sequenceA $
[
-- Library sources.
fmap concat
. withAllLib $ \Library { exposedModules = modules, libBuildInfo = libBi } ->
allSourcesBuildInfo libBi pps modules
-- Executables sources.
, fmap concat
. withAllExe $ \Executable { modulePath = mainPath, buildInfo = exeBi } -> do
biSrcs <- allSourcesBuildInfo exeBi pps []
mainSrc <- findMainExeFile exeBi pps mainPath
return (mainSrc:biSrcs)
-- Test suites sources.
, fmap concat
. withAllTest $ \t -> do
let bi = testBuildInfo t
case testInterface t of
TestSuiteExeV10 _ mainPath -> do
biSrcs <- allSourcesBuildInfo bi pps []
srcMainFile <- do
ppFile <- findFileWithExtension (ppSuffixes pps)
(hsSourceDirs bi) (dropExtension mainPath)
case ppFile of
Nothing -> findFile (hsSourceDirs bi) mainPath
Just pp -> return pp
return (srcMainFile:biSrcs)
TestSuiteLibV09 _ m ->
allSourcesBuildInfo bi pps [m]
TestSuiteUnsupported tp -> die $ "Unsupported test suite type: "
++ show tp
-- Benchmarks sources.
, fmap concat
. withAllBenchmark $ \bm -> do
let bi = benchmarkBuildInfo bm
case benchmarkInterface bm of
BenchmarkExeV10 _ mainPath -> do
biSrcs <- allSourcesBuildInfo bi pps []
srcMainFile <- do
ppFile <- findFileWithExtension (ppSuffixes pps)
(hsSourceDirs bi) (dropExtension mainPath)
case ppFile of
Nothing -> findFile (hsSourceDirs bi) mainPath
Just pp -> return pp
return (srcMainFile:biSrcs)
BenchmarkUnsupported tp -> die $ "Unsupported benchmark type: "
++ show tp
-- Data files.
, fmap concat
. for (dataFiles pkg_descr) $ \filename ->
matchFileGlob (dataDir pkg_descr </> filename)
-- Extra doc files.
, fmap concat
. for (extraDocFiles pkg_descr) $ \ filename ->
matchFileGlob filename
-- License file(s).
, return (licenseFiles pkg_descr)
-- Install-include files.
, fmap concat
. withAllLib $ \ l -> do
let lbi = libBuildInfo l
relincdirs = "." : filter isRelative (includeDirs lbi)
traverse (fmap snd . findIncludeFile relincdirs) (installIncludes lbi)
-- Setup script, if it exists.
, fmap (maybe [] (\f -> [f])) $ findSetupFile ""
-- The .cabal file itself.
, fmap (\d -> [d]) (defaultPackageDesc verbosity)
]
where
-- We have to deal with all libs and executables, so we have local
-- versions of these functions that ignore the 'buildable' attribute:
withAllLib action = traverse action (allLibraries pkg_descr)
withAllExe action = traverse action (executables pkg_descr)
withAllTest action = traverse action (testSuites pkg_descr)
withAllBenchmark action = traverse action (benchmarks pkg_descr)
-- |Prepare a directory tree of source files.
prepareTree :: Verbosity -- ^verbosity
-> PackageDescription -- ^info from the cabal file
-> Maybe LocalBuildInfo
-> FilePath -- ^source tree to populate
-> [PPSuffixHandler] -- ^extra preprocessors (includes suffixes)
-> IO ()
prepareTree verbosity pkg_descr0 mb_lbi targetDir pps = do
-- If the package was configured then we can run platform-independent
-- pre-processors and include those generated files.
case mb_lbi of
Just lbi | not (null pps) -> do
let lbi' = lbi{ buildDir = targetDir </> buildDir lbi }
withAllComponentsInBuildOrder pkg_descr lbi' $ \c clbi ->
preprocessComponent pkg_descr c lbi' clbi True verbosity pps
_ -> return ()
(ordinary, mExecutable) <- listPackageSources verbosity pkg_descr0 pps
installOrdinaryFiles verbosity targetDir (zip (repeat []) ordinary)
installMaybeExecutableFiles verbosity targetDir (zip (repeat []) mExecutable)
maybeCreateDefaultSetupScript targetDir
where
pkg_descr = filterAutogenModules pkg_descr0
-- | Find the setup script file, if it exists.
findSetupFile :: FilePath -> NoCallStackIO (Maybe FilePath)
findSetupFile targetDir = do
hsExists <- doesFileExist setupHs
lhsExists <- doesFileExist setupLhs
if hsExists
then return (Just setupHs)
else if lhsExists
then return (Just setupLhs)
else return Nothing
where
setupHs = targetDir </> "Setup.hs"
setupLhs = targetDir </> "Setup.lhs"
-- | Create a default setup script in the target directory, if it doesn't exist.
maybeCreateDefaultSetupScript :: FilePath -> NoCallStackIO ()
maybeCreateDefaultSetupScript targetDir = do
mSetupFile <- findSetupFile targetDir
case mSetupFile of
Just _setupFile -> return ()
Nothing -> do
writeUTF8File (targetDir </> "Setup.hs") $ unlines [
"import Distribution.Simple",
"main = defaultMain"]
-- | Find the main executable file.
findMainExeFile :: BuildInfo -> [PPSuffixHandler] -> FilePath -> IO FilePath
findMainExeFile exeBi pps mainPath = do
ppFile <- findFileWithExtension (ppSuffixes pps) (hsSourceDirs exeBi)
(dropExtension mainPath)
case ppFile of
Nothing -> findFile (hsSourceDirs exeBi) mainPath
Just pp -> return pp
-- | Given a list of include paths, try to find the include file named
-- @f@. Return the name of the file and the full path, or exit with error if
-- there's no such file.
findIncludeFile :: [FilePath] -> String -> IO (String, FilePath)
findIncludeFile [] f = die ("can't find include file " ++ f)
findIncludeFile (d:ds) f = do
let path = (d </> f)
b <- doesFileExist path
if b then return (f,path) else findIncludeFile ds f
-- | Remove the auto-generated modules (like 'Paths_*') from 'exposed-modules'
-- and 'other-modules'.
filterAutogenModules :: PackageDescription -> PackageDescription
filterAutogenModules pkg_descr0 = mapLib filterAutogenModuleLib $
mapAllBuildInfo filterAutogenModuleBI pkg_descr0
where
mapLib f pkg = pkg { library = fmap f (library pkg)
, subLibraries = map f (subLibraries pkg) }
filterAutogenModuleLib lib = lib {
exposedModules = filter (filterFunction (libBuildInfo lib)) (exposedModules lib)
}
filterAutogenModuleBI bi = bi {
otherModules = filter (filterFunction bi) (otherModules bi)
}
pathsModule = autogenPathsModuleName pkg_descr0
filterFunction bi = \mn ->
mn /= pathsModule
&& not (elem mn (autogenModules bi))
-- | Prepare a directory tree of source files for a snapshot version.
-- It is expected that the appropriate snapshot version has already been set
-- in the package description, eg using 'snapshotPackage' or 'snapshotVersion'.
--
prepareSnapshotTree :: Verbosity -- ^verbosity
-> PackageDescription -- ^info from the cabal file
-> Maybe LocalBuildInfo
-> FilePath -- ^source tree to populate
-> [PPSuffixHandler] -- ^extra preprocessors (includes
-- suffixes)
-> IO ()
prepareSnapshotTree verbosity pkg mb_lbi targetDir pps = do
prepareTree verbosity pkg mb_lbi targetDir pps
overwriteSnapshotPackageDesc verbosity pkg targetDir
overwriteSnapshotPackageDesc :: Verbosity -- ^verbosity
-> PackageDescription -- ^info from the cabal file
-> FilePath -- ^source tree
-> IO ()
overwriteSnapshotPackageDesc verbosity pkg targetDir = do
-- We could just writePackageDescription targetDescFile pkg_descr,
-- but that would lose comments and formatting.
descFile <- defaultPackageDesc verbosity
withUTF8FileContents descFile $
writeUTF8File (targetDir </> descFile)
. unlines . map (replaceVersion (packageVersion pkg)) . lines
where
replaceVersion :: Version -> String -> String
replaceVersion version line
| "version:" `isPrefixOf` map toLower line
= "version: " ++ display version
| otherwise = line
-- | Modifies a 'PackageDescription' by appending a snapshot number
-- corresponding to the given date.
--
snapshotPackage :: UTCTime -> PackageDescription -> PackageDescription
snapshotPackage date pkg =
pkg {
package = pkgid { pkgVersion = snapshotVersion date (pkgVersion pkgid) }
}
where pkgid = packageId pkg
-- | Modifies a 'Version' by appending a snapshot number corresponding
-- to the given date.
--
snapshotVersion :: UTCTime -> Version -> Version
snapshotVersion date version = version {
versionBranch = versionBranch version
++ [dateToSnapshotNumber date]
}
-- | Given a date produce a corresponding integer representation.
-- For example given a date @18/03/2008@ produce the number @20080318@.
--
dateToSnapshotNumber :: UTCTime -> Int
dateToSnapshotNumber date = case toGregorian (utctDay date) of
(year, month, day) ->
fromIntegral year * 10000
+ month * 100
+ day
-- | Callback type for use by sdistWith.
type CreateArchiveFun = Verbosity -- ^verbosity
-> PackageDescription -- ^info from cabal file
-> Maybe LocalBuildInfo -- ^info from configure
-> FilePath -- ^source tree to archive
-> FilePath -- ^name of archive to create
-> IO FilePath
-- | Create an archive from a tree of source files, and clean up the tree.
createArchive :: CreateArchiveFun
createArchive verbosity pkg_descr mb_lbi tmpDir targetPref = do
let tarBallFilePath = targetPref </> tarBallName pkg_descr <.> "tar.gz"
(tarProg, _) <- requireProgram verbosity tarProgram
(maybe defaultProgramDb withPrograms mb_lbi)
let formatOptSupported = maybe False (== "YES") $
Map.lookup "Supports --format"
(programProperties tarProg)
runProgram verbosity tarProg $
-- Hmm: I could well be skating on thinner ice here by using the -C option
-- (=> seems to be supported at least by GNU and *BSD tar) [The
-- prev. solution used pipes and sub-command sequences to set up the paths
-- correctly, which is problematic in a Windows setting.]
["-czf", tarBallFilePath, "-C", tmpDir]
++ (if formatOptSupported then ["--format", "ustar"] else [])
++ [tarBallName pkg_descr]
return tarBallFilePath
-- | Given a buildinfo, return the names of all source files.
allSourcesBuildInfo :: BuildInfo
-> [PPSuffixHandler] -- ^ Extra preprocessors
-> [ModuleName] -- ^ Exposed modules
-> IO [FilePath]
allSourcesBuildInfo bi pps modules = do
let searchDirs = hsSourceDirs bi
sources <- fmap concat $ sequenceA $
[ let file = ModuleName.toFilePath module_
in findAllFilesWithExtension suffixes searchDirs file
>>= nonEmpty (notFound module_) return
| module_ <- modules ++ otherModules bi ]
bootFiles <- sequenceA
[ let file = ModuleName.toFilePath module_
fileExts = ["hs-boot", "lhs-boot"]
in findFileWithExtension fileExts (hsSourceDirs bi) file
| module_ <- modules ++ otherModules bi ]
return $ sources ++ catMaybes bootFiles ++ cSources bi ++ jsSources bi
where
nonEmpty x _ [] = x
nonEmpty _ f xs = f xs
suffixes = ppSuffixes pps ++ ["hs", "lhs"]
notFound m = die $ "Error: Could not find module: " ++ display m
++ " with any suffix: " ++ show suffixes ++ ". If the module "
++ "is autogenerated it should be added to 'autogen-modules'."
printPackageProblems :: Verbosity -> PackageDescription -> IO ()
printPackageProblems verbosity pkg_descr = do
ioChecks <- checkPackageFiles pkg_descr "."
let pureChecks = checkConfiguredPackage pkg_descr
isDistError (PackageDistSuspicious _) = False
isDistError (PackageDistSuspiciousWarn _) = False
isDistError _ = True
(errors, warnings) = partition isDistError (pureChecks ++ ioChecks)
unless (null errors) $
notice verbosity $ "Distribution quality errors:\n"
++ unlines (map explanation errors)
unless (null warnings) $
notice verbosity $ "Distribution quality warnings:\n"
++ unlines (map explanation warnings)
unless (null errors) $
notice verbosity
"Note: the public hackage server would reject this package."
------------------------------------------------------------
-- | The name of the tarball without extension
--
tarBallName :: PackageDescription -> String
tarBallName = display . packageId
mapAllBuildInfo :: (BuildInfo -> BuildInfo)
-> (PackageDescription -> PackageDescription)
mapAllBuildInfo f pkg = pkg {
library = fmap mapLibBi (library pkg),
subLibraries = fmap mapLibBi (subLibraries pkg),
executables = fmap mapExeBi (executables pkg),
testSuites = fmap mapTestBi (testSuites pkg),
benchmarks = fmap mapBenchBi (benchmarks pkg)
}
where
mapLibBi lib = lib { libBuildInfo = f (libBuildInfo lib) }
mapExeBi exe = exe { buildInfo = f (buildInfo exe) }
mapTestBi t = t { testBuildInfo = f (testBuildInfo t) }
mapBenchBi bm = bm { benchmarkBuildInfo = f (benchmarkBuildInfo bm) }
| sopvop/cabal | Cabal/Distribution/Simple/SrcDist.hs | bsd-3-clause | 19,745 | 0 | 24 | 5,249 | 3,999 | 2,037 | 1,962 | 326 | 6 |
{-# LANGUAGE Rank2Types #-}
module Examples
( tests
) where
import Control.Applicative ((<$>))
import Control.Monad (forM)
import Data.Bits (shiftL)
import Data.List (isPrefixOf, sort)
import Test.Framework (Test, testGroup)
import Test.Framework.Providers.HUnit (testCase)
import Test.HUnit (Assertion, assert, (@=?))
import qualified Data.ByteString as B
import Assembler
import Emulator
import Emulator.Monad
import Emulator.Monad.ST
import Memory (Address (..), Register (..))
tests :: Test
tests = testGroup "Examples"
[ testExample "notch" $ do
x <- load $ Register X
cycles <- load Cycles
return $ (0x40, 106) @=? (x, cycles)
, testExample "sum-squares" $ do
x <- load $ Register X
return $ sum [n * n | n <- [0 .. 50]] @=? x
, testExample "bubble-sort" $ do
xs <- forM [0 .. 9] $ load . Ram . (0x1000 +)
return $ sort xs @=? xs
, testExample "32-bit-add" $ do
lo <- load $ Ram 0x1000
hi <- load $ Ram 0x1001
let sum' = (fromIntegral hi `shiftL` 16) + fromIntegral lo :: Int
return $ 0x12345678 + 0xaabbccdd @=? sum'
, testExample "fib" $ do
let fibs = 1 : 2 : zipWith (+) fibs (tail fibs)
addrs = [0xffff, 0xfffe .. 0x000c]
loop _ [] = return True
loop [] _ = return True
loop (f : fs) (a : as) = do
f' <- load $ Ram a
if f == f' then loop fs as else return False
return . assert =<< loop fibs addrs
, testExample "self-copy" $ do
let readRam i = do
x <- load $ Ram i
if x == 0x0000
then return []
else (x :) <$> readRam (i + 1)
programs <- readRam 1
let len = length programs `div` 10
equal xs = case splitAt len xs of
(_, []) -> True
(hs, ts) -> hs `isPrefixOf` ts && equal ts
return $ assert $ equal programs
]
testExample :: String
-> (forall s. STEmulator s Assertion)
-> Test
testExample name = testCase name .
example ("examples/" ++ name ++ ".dasm16")
example :: FilePath
-> (forall s. STEmulator s Assertion)
-> Assertion
example filePath check = do
assembleFile filePath "a.out"
program <- B.readFile "a.out"
runSTEmulator $ do
loadProgram program
emulate
check
| jaspervdj/dcpu16-hs | tests/Examples.hs | bsd-3-clause | 2,482 | 0 | 19 | 861 | 878 | 454 | 424 | 69 | 6 |
{-# LANGUAGE FlexibleInstances #-}
{-# LANGUAGE MultiParamTypeClasses #-}
{-# LANGUAGE RankNTypes #-}
{-# LANGUAGE GADTs #-}
module Oracle.DiffOracle where
import Data.Maybe
import Data.Tuple (swap)
import Debug.Trace
import Control.Monad.Reader
import qualified Data.IntMap as M
import qualified Data.Set as S
import Oracle.Internal
import Language.Common
import Language.Clojure.AST
import Language.Clojure.Lang
import Util.UnixDiff
type CopyMap = M.IntMap Int
type CopyMaps = (CopyMap, CopyMap)
type MbMoveConflict = ConflictResult (S.Set (Int, Int))
type DelInsMap = (M.IntMap Path, M.IntMap Path)
data DiffOracle = DiffOracle DelInsMap
deriving (Show)
buildDiffOracle :: String -> String -> Expr -> Expr -> DiffOracle
buildDiffOracle s d src dst = DiffOracle diffActions
where
cp = buildCopyMaps (preprocess s d)
delinsMap = buildDelInsMap (preprocessGrouped s d)
diffActions = solveConflicts delinsMap cp src dst
buildCopyMaps :: [DiffAction] -> (M.IntMap Int, M.IntMap Int)
buildCopyMaps as = (insMap, reverseMap insMap)
where
insMap = buildCopyMap as
reverseMap = M.fromList . map swap . M.toList
buildCopyMap [] = (M.empty)
buildCopyMap (first:rest) = (process first) `M.union` (buildCopyMap rest)
where
process (Copy (i1, i2)) = M.singleton i1 i2
process _ = M.empty
solveConflicts :: DelInsMap -> CopyMaps -> Expr -> Expr -> DelInsMap
solveConflicts diffActions copyMaps src dst = foldl invalidate diffActions conflicts
where
conflicts = checkCopyMaps copyMaps src dst
invalidate :: DelInsMap -> [MbMoveConflict] -> DelInsMap
invalidate diffActions [] = diffActions
invalidate diffActions (NoConflict:rest) = invalidate diffActions rest
invalidate diffActions ((ConflictAt pairs):rest) = invalidate (S.foldl invalidatePair diffActions pairs) rest
invalidatePair :: DelInsMap -> (Int, Int) -> DelInsMap
invalidatePair (srcMap, dstMap) (i,j) = (M.insert i M srcMap, M.insert j M dstMap)
unionDelInsMap :: DelInsMap -> DelInsMap -> DelInsMap
unionDelInsMap (s1, d1) (s2, d2) = (M.union s1 s2, M.union d1 d2)
buildDelInsMap :: [GroupDiffAction] -> DelInsMap
buildDelInsMap [] = (M.empty, M.empty)
buildDelInsMap (first:rest) = process first `unionDelInsMap` buildDelInsMap rest
where
process (OMod srcRange dstRange) = (insertRange srcRange M, insertRange dstRange M)
process (OIns dstRange _) = (M.empty, insertRange dstRange I)
process (ODel srcRange _) = (insertRange srcRange D, M.empty)
insertRange :: LineRange -> Path -> M.IntMap Path
insertRange (Range s e) o = go s M.empty
where
go i m | i <= e = go (i+1) (M.insert i o m)
| otherwise = m
giveAdvice' :: DelInsMap -> Usingl u -> Usingl v -> [Path]
giveAdvice' (srcMap, dstMap) src dst =
if (isMod srcRange srcMap && isMod dstRange dstMap)
then []
else if (isDel srcRange srcMap || isMod srcRange srcMap)
then [ D ]
else if (isIns dstRange dstMap || isMod dstRange dstMap)
then [ I ]
else [ M ]
where
srcRange = fromJust $ extractRange src
dstRange = fromJust $ extractRange dst
isContainedIn :: LineRange -> LineRange -> Bool
isContainedIn (Range s1 e1) (Range s2 e2) = s2 <= s1 && e1 <= e2
isMod :: LineRange -> M.IntMap Path -> Bool
isMod lr m = case M.lookup (takeStart lr) m of
Just M -> True
_ -> False
isIns :: LineRange -> M.IntMap Path -> Bool
isIns lr m = case M.lookup (takeStart lr) m of
Just I -> True
_ -> False
isDel :: LineRange -> M.IntMap Path -> Bool
isDel lr m = case M.lookup (takeStart lr) m of
Just D -> True
_ -> False
inSync :: Usingl u -> LineRange -> Bool
inSync u lr = uRange `inSync'` lr
where
uRange = fromJust $ extractRange u
inSync' :: LineRange -> LineRange -> Bool
inSync' (Range s1 e1) (Range s2 e2) = s2 <= s1 && s1 <= e2
copySetExpr :: CopyMap -> Expr -> S.Set Int
copySetExpr copyMap e = collectAll copyMap eRange
where
eRange = extractRangeExpr e
copySetSel :: CopyMap -> SepExprList -> S.Set Int
copySetSel copyMap e = collectAll copyMap eRange
where
eRange = extractRangeSepExprList e
collectAll :: CopyMap -> LineRange -> S.Set Int
collectAll cpM (Range s e) = go cpM s
where
go m i | i <= e = mbTakeLine m i `S.union` go m (i+1)
go m i | otherwise = S.empty
mbTakeLine m i = if isJust (M.lookup i m)
then S.singleton i
else S.empty
lookupSet :: S.Set Int -> CopyMap -> S.Set Int
lookupSet s cp = S.map (fromJust . flip M.lookup cp) s
lookupSetPairs :: S.Set Int -> CopyMap -> [(Int, Int)]
lookupSetPairs s cp = S.toList $ S.map (\i -> (i, fromJust (M.lookup i cp))) s
intersectsNonOverlapping :: Ord a => S.Set a -> S.Set a -> S.Set a -> S.Set a
intersectsNonOverlapping target a b =
if check
then target `S.difference` overlapping
else S.empty
where
check = target `intersects` (a `S.difference` overlapping) && target `intersects` (b `S.difference` overlapping)
overlapping = a `S.intersection` b
targetL = S.toList target
intersects :: Ord a => S.Set a -> S.Set a -> Bool
intersects a b = not (S.null (a `S.intersection` b))
pickBigger :: CopyMaps -> S.Set Int -> S.Set Int -> S.Set (Int, Int)
pickBigger (srcMap, dstMap) a b
| S.size a >= S.size b = S.fromList $ map swap (lookupSetPairs a dstMap)
| otherwise = S.fromList $ lookupSetPairs b srcMap
deOptimizeExpr :: CopyMaps -> Expr -> Expr -> MbMoveConflict
deOptimizeExpr cp@(srcMap, dstMap) (Seq a b _) (Seq c d _) =
if (S.null overlapA && S.null overlapC)
then NoConflict
else ConflictAt (pickBigger cp overlapA overlapC)
where
copySetA = copySetExpr srcMap a
copySetB = copySetExpr srcMap b
copySetC = copySetExpr dstMap c
copySetD = copySetExpr dstMap d
copyTargetA = lookupSet copySetA srcMap
copyTargetC = lookupSet copySetC dstMap
overlapA = intersectsNonOverlapping copyTargetA copySetC copySetD
overlapC = intersectsNonOverlapping copyTargetC copySetA copySetB
deOptimizeExpr cp@(srcMap, dstMap) (Collection _ (Cons a _ b _) _) (Collection _ (Cons c _ d _) _) =
if (S.null overlapA && S.null overlapC)
then NoConflict
else ConflictAt (pickBigger cp overlapA overlapC)
where
copySetA = copySetExpr srcMap a
copySetB = copySetSel srcMap b
copySetC = copySetExpr dstMap c
copySetD = copySetSel dstMap d
copyTargetA = lookupSet copySetA srcMap
copyTargetC = lookupSet copySetC dstMap
overlapA = intersectsNonOverlapping copyTargetA copySetC copySetD
overlapC = intersectsNonOverlapping copyTargetC copySetA copySetB
deOptimizeExpr _ _ _ = NoConflict
checkCopyMaps :: CopyMaps -> Expr -> Expr -> [[MbMoveConflict]]
checkCopyMaps cp@(srcMap, dstMap) src dst = fmap (collectSrcDstLines cp src dst) (M.toList srcMap)
collectSrcDstLines :: CopyMaps -> Expr -> Expr -> (Int, Int) -> [MbMoveConflict]
collectSrcDstLines cp src dst (s,d) = map (\(src,dst) -> deOptimizeExpr cp (wrap src) (wrap dst)) conflifts
where
conflifts = zipEqLen srcConflicts dstConflicts
srcConflicts = collectSubTrees src s
dstConflicts = collectSubTrees dst d
zipEqLen :: [a] -> [b] -> [(a,b)]
zipEqLen (a:as) (b:bs) = (a,b):(zipEqLen as bs)
zipEqLen [] [] = []
zipEqLen [] bs = error "dst is longer"
zipEqLen as [] = error "src is longer"
wrap :: SubTree -> Expr
wrap (Exp e) = e
wrap (Sel sel) = (Collection Parens sel (extractRangeSepExprList sel))
instance (Monad m) => OracleF DiffOracle m where
callF o@(DiffOracle diffActions) s d = return $ askOracle o s d
instance (Monad m) => OracleP DiffOracle m where
callP _ An An = do
return []
callP _ An (_ `Ac` _) = do
return [ I ]
callP _ (_ `Ac` _) An = do
return [ D ]
callP o@(DiffOracle diffActions) (s `Ac` _) (d `Ac` _) = return $ askOracle o s d
askOracle :: DiffOracle -> Usingl u -> Usingl v -> [Path]
askOracle (DiffOracle diffActions) src dst = case (extractRange src, extractRange dst) of
(Nothing, Nothing) -> [ M ]
(Just sRange, Nothing) -> [ D ]
(Nothing, Just dRange) -> [ I ]
(Just sRange, Just dRange) -> giveAdvice' diffActions src dst | nazrhom/vcs-clojure | src/Oracle/DiffOracle.hs | bsd-3-clause | 8,201 | 0 | 13 | 1,776 | 3,167 | 1,644 | 1,523 | 174 | 4 |
{-# LANGUAGE
DeriveFunctor
, FlexibleInstances
, ScopedTypeVariables
#-}
module Data.Trie.Pseudo where
import Prelude hiding (foldl, foldr, foldr1, lookup,
map)
import Data.Foldable hiding (all)
import Data.List (intercalate)
import Data.List.NonEmpty (NonEmpty (..), fromList, toList)
import qualified Data.List.NonEmpty as NE
import Data.Maybe (fromMaybe)
import Data.Monoid
import qualified Data.Semigroup as S
import Control.Applicative
import Control.Monad (replicateM)
import Control.Arrow (second)
-- TODO: difference
-- | Tagged rose tree with explicit emptyness
data PseudoTrie t a = More t (Maybe a) (NonEmpty (PseudoTrie t a))
| Rest (NonEmpty t) a
| Nil
deriving (Show, Eq, Functor)
-- | Overwriting instance
instance (Eq t) => Monoid (PseudoTrie t a) where
mempty = Nil
mappend = merge
-- | Depth first
instance Foldable (PseudoTrie t) where
foldr _ acc Nil = acc
foldr f acc (Rest _ x) = f x acc
foldr f acc (More t Nothing xs) = foldr go acc xs
where
go z bcc = foldr f bcc z
foldr f acc (More t (Just x) xs) = foldr go (f x acc) xs
where
go z bcc = foldr f bcc z
beginsWith :: (Eq t) => PseudoTrie t a -> t -> Bool
beginsWith Nil _ = False
beginsWith (Rest (t:|_) _) p = t == p
beginsWith (More t _ _) p = t == p
-- | Provides a form of deletion by setting a path to @Nothing@, but doesn't
-- cleanup like @prune@
assign :: (Eq t) => NonEmpty t -> Maybe a -> PseudoTrie t a -> PseudoTrie t a
assign ts (Just x) Nil = Rest ts x
assign _ Nothing Nil = Nil
assign tss@(t:|ts) mx ys@(Rest pss@(p:|ps) y)
| tss == pss = case mx of
(Just x) -> Rest pss x
Nothing -> Nil
| t == p = case (ts,ps) of
([], p':_) -> More t mx $ Rest (NE.fromList ps) y :| []
(t':_, []) -> case mx of
Just x -> More p (Just y) $ Rest (NE.fromList ts) x :| []
Nothing -> ys
(t':_,p':_) -> if t' == p'
then More t Nothing $
assign (NE.fromList ts) mx (Rest (NE.fromList ps) y) :| []
else case mx of -- disjoint
Nothing -> ys
Just x -> More t Nothing $ NE.fromList $
[ Rest (NE.fromList ps) y
, Rest (NE.fromList ts) x
]
| otherwise = ys
assign (t:|ts) mx y@(More p my ys)
| t == p = case ts of
[] -> More p mx ys
_ -> More p my $ fmap (assign (NE.fromList ts) mx) ys
| otherwise = y
-- | Overwrite the LHS point-wise with the RHS's contents
merge :: (Eq t) => PseudoTrie t a -> PseudoTrie t a -> PseudoTrie t a
merge Nil y = y
merge x Nil = x
merge xx@(Rest tss@(t:|ts) x) (Rest pss@(p:|ps) y)
| tss == pss = Rest pss y
| t == p = case (ts,ps) of
([],p':ps') -> More t (Just x) $ Rest (NE.fromList ps) y :| []
(t':ts',[]) -> More t (Just y) $ Rest (NE.fromList ts) x :| []
(_,_) -> More t Nothing $
merge (Rest (NE.fromList ts) x)
(Rest (NE.fromList ps) y) :| []
| otherwise = xx
merge xx@(More t mx xs) (More p my ys)
| t == p = More p my $ NE.fromList $
foldr go [] $ NE.toList xs ++ NE.toList ys
| otherwise = xx
where
go q [] = [q]
go q (z:zs) | areDisjoint q z = q : z : zs
| otherwise = merge q z : zs
merge xx@(More t mx xs) (Rest pss@(p:|ps) y)
| t == p = case ps of
[] -> More t (Just y) xs
_ -> More t mx $
fmap (flip merge $ Rest (NE.fromList ps) y) xs
| otherwise = xx
merge xx@(Rest tss@(t:|ts) x) (More p my ys)
| t == p = case ts of
[] -> More p (Just x) ys
_ -> More p my $
fmap (merge $ Rest (NE.fromList ts) x) ys
| otherwise = xx
add :: (Eq t) => NonEmpty t -> PseudoTrie t a -> PseudoTrie t a -> PseudoTrie t a
add ts input container =
let ts' = NE.toList ts in
merge container $ mkMores ts' input
where
mkMores :: (Eq t) => [t] -> PseudoTrie t a -> PseudoTrie t a
mkMores [] trie = trie
mkMores (t:ts) trie = More t Nothing $
mkMores ts trie :| []
toAssocs :: PseudoTrie t a -> [(NonEmpty t, a)]
toAssocs = go [] []
where
go :: [t] -> [(NonEmpty t, a)] -> PseudoTrie t a -> [(NonEmpty t, a)]
go depth acc Nil = acc
go depth acc (Rest ts x) = (NE.fromList $ depth ++ NE.toList ts, x) : acc
go depth acc (More t Nothing xs) =
foldr (flip $ go $ depth ++ [t]) acc $ NE.toList xs
go depth acc (More t (Just x) xs) =
(NE.fromList $ depth ++ [t], x) :
(foldr $ flip $ go $ depth ++ [t]) acc (NE.toList xs)
fromAssocs :: (Eq t) => [(NonEmpty t, a)] -> PseudoTrie t a
fromAssocs = foldr (uncurry assign) Nil . fmap (second Just)
lookup :: (Eq t) => NonEmpty t -> PseudoTrie t a -> Maybe a
lookup _ Nil = Nothing
lookup tss (Rest pss a)
| tss == pss = Just a
| otherwise = Nothing
lookup tss@(t:|ts) (More p mx xs)
| t == p = case ts of
[] -> mx
(t':ts') -> find (hasNextTag t') xs >>= lookup (fromList ts)
| otherwise = Nothing
where
hasNextTag :: (Eq t) => t -> PseudoTrie t a -> Bool
hasNextTag t Nil = False
hasNextTag t (More p _ _) = t == p
hasNextTag t (Rest (p:|_) _) = t == p
-- | Simple test on the heads of two tries
areDisjoint :: (Eq t) => PseudoTrie t a -> PseudoTrie t a -> Bool
areDisjoint (More t _ _) (More p _ _)
| t == p = False
| otherwise = True
areDisjoint (Rest (t:|_) _) (Rest (p:|_) _)
| t == p = False
| otherwise = True
areDisjoint _ _ = True
-- | The meet of two @PseudoTrie@s
intersectionWith :: (Eq t) =>
(a -> b -> c)
-> PseudoTrie t a
-> PseudoTrie t b
-> PseudoTrie t c
intersectionWith _ _ Nil = Nil
intersectionWith _ Nil _ = Nil
intersectionWith f (Rest tss@(t:|ts) x) (Rest pss@(p:|ps) y)
| tss == pss = Rest pss $ f x y
| otherwise = Nil
intersectionWith f (More t mx xs) (More p my ys)
| t == p = case [intersectionWith f x' y' | x' <- NE.toList xs, y' <- NE.toList ys] of
[] -> case f <$> mx <*> my of
Nothing -> Nil
Just c -> Rest (p :| []) c
zs -> More p (f <$> mx <*> my) $ NE.fromList zs
-- implicit root
| otherwise = Nil
intersectionWith f (More t mx xs) (Rest pss@(p:|ps) y)
| t == p = case ps of
[] -> case f <$> mx <*> Just y of
Nothing -> Nil
Just c -> Rest (p :| []) c
_ -> More p Nothing $ fmap (flip (intersectionWith f) $ Rest (fromList ps) y) xs
| otherwise = Nil
intersectionWith f (Rest tss@(t:|ts) x) (More p my ys)
| t == p = case ts of
[] -> case f <$> Just x <*> my of
Nothing -> Nil
Just c -> Rest (t :| []) c
_ -> More t Nothing $ fmap (intersectionWith f $ Rest (fromList ts) x) ys
| otherwise = Nil
-- difference :: Eq t =>
-- PseudoTrie t a
-- -> PseudoTrie t a
-- -> PseudoTrie t a
-- | Needless intermediary elements are turned into shortcuts, @Nil@'s in
-- subtrees are also removed.
prune :: PseudoTrie t a -> PseudoTrie t a
prune = go
where
go Nil = Nil
go xx@(Rest ts x) = xx
go (More t Nothing xs) =
case cleaned xs of
[Nil] -> Nil
[Rest ts x] -> Rest (t:|NE.toList ts) x
xs' -> More t Nothing $ NE.fromList xs'
go (More t (Just x) xs) =
case cleaned xs of
[Nil] -> Rest (t:|[]) x
xs' -> More t (Just x) $ NE.fromList xs'
cleaned xs = removeNils (NE.toList $ fmap go xs)
removeNils xs = case removeNils' xs of
[] -> [Nil]
ys -> ys
where
removeNils' [] = []
removeNils' (Nil:xs) = removeNils' xs
removeNils' (x:xs) = x : removeNils' xs
| athanclark/pseudo-trie | src/Data/Trie/Pseudo.hs | bsd-3-clause | 8,552 | 0 | 18 | 3,279 | 3,708 | 1,862 | 1,846 | 185 | 10 |
-- | Test utility functions
module Test.Util(utilsTests) where
import Test.Tasty
import Test.Tasty.HUnit
import Language.Haskell.Ghcid.Util
utilsTests :: TestTree
utilsTests = testGroup "Utility tests"
[dropPrefixTests
,chunksOfWordTests
]
dropPrefixTests :: TestTree
dropPrefixTests = testGroup "dropPrefix"
[testCase "Prefix not found" $ dropPrefixRepeatedly "prefix" "string" @?= "string"
,testCase "Empty prefix" $ dropPrefixRepeatedly "" "string" @?= "string"
,testCase "Prefix found once" $ dropPrefixRepeatedly "str" "string" @?= "ing"
,testCase "Prefix found twice" $ dropPrefixRepeatedly "str" "strstring" @?= "ing"
]
chunksOfWordTests :: TestTree
chunksOfWordTests = testGroup "chunksOfWord"
[testCase "Max 0" $ chunksOfWord 4 0 "ab cd efgh" @?= ["ab c","d ef","gh"]
,testCase "Max 2" $ chunksOfWord 4 2 "ab cd efgh" @?= ["ab ","cd ","efgh"]
]
| JPMoresmau/ghcid | src/Test/Util.hs | bsd-3-clause | 904 | 0 | 9 | 158 | 217 | 116 | 101 | 18 | 1 |
{-# LANGUAGE TemplateHaskell #-}
{-# LANGUAGE ScopedTypeVariables #-}
{-# LANGUAGE OverloadedStrings #-}
{-# LANGUAGE RecordWildCards #-}
module TauSigma.ADEV
( Statistic(..)
, Options
, options
, main
) where
import Control.Monad.Primitive (PrimMonad)
import Control.Monad.Trans
import Control.Monad.Trans.Except
import Control.Lens (view)
import Control.Lens.TH
import Control.Parallel.Strategies (withStrategy, parBuffer, rdeepseq)
import Data.Csv (HasHeader(..), fromOnly)
import qualified Data.Vector as V
import qualified Data.Vector.Unboxed as U
import Options.Applicative hiding (header)
import Pipes
import Pipes.ByteString (stdin, stdout)
import qualified Pipes.Prelude as P
import TauSigma.Types (TauSigma(..))
import TauSigma.Statistics.Types (Tau0, Tau, Sigma)
import TauSigma.Statistics.Allan (adevs, mdevs, tdevs)
import TauSigma.Statistics.Hadamard (hdevs)
import TauSigma.Statistics.Total (totdevs)
import TauSigma.Statistics.Theo1 (theo1devs, theoBRdevs, theoHdevs)
import TauSigma.Util.CSV
import TauSigma.Util.Vector (drainToVector)
data Statistic = ADEV | MDEV | TDEV | HDEV | TOTDEV | Theo1 | TheoBR | TheoH
data Options
= Options { _tau0 :: Tau0 Double
, _maxTau :: Maybe (Tau Double)
}
$(makeLenses ''Options)
options :: Parser Options
options = Options <$> tau0 <*> maxTau
where f `with` xs = f (mconcat xs)
tau0 = option auto
`with` [ long "tau0"
, metavar "N"
, help "Base sampling interval"
]
maxTau = option (fmap Just auto)
`with` [ long "max-tau"
, metavar "N"
, value Nothing
, help "Maximum multiple of tau0 to output."
]
main :: (PrimMonad m, MonadIO m) =>
Statistic
-> Options
-> ExceptT String m ()
main statistic opts = do
errors <- drainToVector (decode NoHeader stdin >-> P.map fromOnly)
runEffect $ each (compute statistic opts errors)
>-> P.map (uncurry TauSigma)
>-> encodeByName (V.fromList ["tau", "sigma"])
>-> stdout
compute
:: Statistic
-> Options
-> U.Vector Double
-> [(Tau Double, Sigma Double)]
compute statistic opts xs = parallelize retained
where parallelize = withStrategy (parBuffer 50 rdeepseq)
retained = limiter opts all
where all = dispatch statistic (view tau0 opts) xs
dispatch
:: Statistic
-> Tau0 Double
-> U.Vector Double
-> [(Tau Double, Sigma Double)]
dispatch ADEV = adevs
dispatch MDEV = mdevs
dispatch TDEV = tdevs
dispatch HDEV = hdevs
dispatch TOTDEV = totdevs
dispatch Theo1 = theo1devs
dispatch TheoBR = theoBRdevs
dispatch TheoH = theoHdevs
limiter
:: Options
-> [(Tau Double, Sigma Double)]
-> [(Tau Double, Sigma Double)]
limiter opts =
case view maxTau opts of
Nothing -> id
Just limit -> limiter' limit
limiter'
:: Tau Double
-> [(Tau Double, Sigma Double)]
-> [(Tau Double, Sigma Double)]
limiter' limit = filter go
where go (tau, _) = tau <= limit
| sacundim/tau-sigma | src/TauSigma/ADEV.hs | bsd-3-clause | 3,151 | 0 | 13 | 812 | 949 | 523 | 426 | 93 | 2 |
{-# LANGUAGE BangPatterns #-}
{-# LANGUAGE CPP #-}
{-# LANGUAGE OverloadedStrings #-}
{-# LANGUAGE PackageImports #-}
{-# LANGUAGE PatternGuards #-}
{-# LANGUAGE ScopedTypeVariables #-}
{-# LANGUAGE TupleSections #-}
{-# OPTIONS_GHC -fno-warn-deprecations #-}
module Network.Wai.Handler.Warp.Run where
#if __GLASGOW_HASKELL__ < 709
import Control.Applicative ((<$>))
#endif
import Control.Arrow (first)
import Control.Concurrent (threadDelay)
import qualified Control.Concurrent as Conc (yield)
import Control.Exception as E
import Control.Monad (when, unless, void)
import Data.ByteString (ByteString)
import qualified Data.ByteString as S
import Data.Char (chr)
import "iproute" Data.IP (toHostAddress, toHostAddress6)
import Data.IORef (IORef, newIORef, readIORef, writeIORef)
import Data.Streaming.Network (bindPortTCP)
import Network (sClose, Socket)
import Network.Socket (accept, withSocketsDo, SockAddr(SockAddrInet, SockAddrInet6), setSocketOption, SocketOption(..))
import qualified Network.Socket.ByteString as Sock
import Network.Wai
import Network.Wai.Handler.Warp.Buffer
import Network.Wai.Handler.Warp.Counter
import qualified Network.Wai.Handler.Warp.Date as D
import qualified Network.Wai.Handler.Warp.FdCache as F
import qualified Network.Wai.Handler.Warp.FileInfoCache as I
import Network.Wai.Handler.Warp.HTTP2 (http2, isHTTP2)
import Network.Wai.Handler.Warp.Header
import Network.Wai.Handler.Warp.ReadInt
import Network.Wai.Handler.Warp.Recv
import Network.Wai.Handler.Warp.Request
import Network.Wai.Handler.Warp.Response
import Network.Wai.Handler.Warp.SendFile
import Network.Wai.Handler.Warp.Settings
import qualified Network.Wai.Handler.Warp.Timeout as T
import Network.Wai.Handler.Warp.Types
import Network.Wai.Internal (ResponseReceived (ResponseReceived))
import System.Environment (getEnvironment)
import System.IO.Error (isFullErrorType, ioeGetErrorType)
#if WINDOWS
import Network.Wai.Handler.Warp.Windows
#else
import Network.Socket (fdSocket)
#endif
-- | Creating 'Connection' for plain HTTP based on a given socket.
socketConnection :: Socket -> IO Connection
socketConnection s = do
bufferPool <- newBufferPool
writeBuf <- allocateBuffer bufferSize
let sendall = Sock.sendAll s
return Connection {
connSendMany = Sock.sendMany s
, connSendAll = sendall
, connSendFile = sendFile s writeBuf bufferSize sendall
, connClose = sClose s >> freeBuffer writeBuf
, connRecv = receive s bufferPool
, connRecvBuf = receiveBuf s
, connWriteBuffer = writeBuf
, connBufferSize = bufferSize
}
#if __GLASGOW_HASKELL__ < 702
allowInterrupt :: IO ()
allowInterrupt = unblock $ return ()
#endif
-- | Run an 'Application' on the given port.
-- This calls 'runSettings' with 'defaultSettings'.
run :: Port -> Application -> IO ()
run p = runSettings defaultSettings { settingsPort = p }
-- | Run an 'Application' on the port present in the @PORT@
-- environment variable. Uses the 'Port' given when the variable is unset.
-- This calls 'runSettings' with 'defaultSettings'.
--
-- Since 3.0.9
runEnv :: Port -> Application -> IO ()
runEnv p app = do
mp <- lookup "PORT" <$> getEnvironment
maybe (run p app) runReadPort mp
where
runReadPort :: String -> IO ()
runReadPort sp = case reads sp of
((p', _):_) -> run p' app
_ -> fail $ "Invalid value in $PORT: " ++ sp
-- | Run an 'Application' with the given 'Settings'.
-- This opens a listen socket on the port defined in 'Settings' and
-- calls 'runSettingsSocket'.
runSettings :: Settings -> Application -> IO ()
runSettings set app = withSocketsDo $
bracket
(bindPortTCP (settingsPort set) (settingsHost set))
sClose
(\socket -> do
setSocketCloseOnExec socket
runSettingsSocket set socket app)
-- | This installs a shutdown handler for the given socket and
-- calls 'runSettingsConnection' with the default connection setup action
-- which handles plain (non-cipher) HTTP.
-- When the listen socket in the second argument is closed, all live
-- connections are gracefully shut down.
--
-- The supplied socket can be a Unix named socket, which
-- can be used when reverse HTTP proxying into your application.
--
-- Note that the 'settingsPort' will still be passed to 'Application's via the
-- 'serverPort' record.
runSettingsSocket :: Settings -> Socket -> Application -> IO ()
runSettingsSocket set socket app = do
settingsInstallShutdownHandler set closeListenSocket
runSettingsConnection set getConn app
where
getConn = do
#if WINDOWS
(s, sa) <- windowsThreadBlockHack $ accept socket
#else
(s, sa) <- accept socket
#endif
setSocketCloseOnExec s
-- NoDelay causes an error for AF_UNIX.
setSocketOption s NoDelay 1 `E.catch` \(E.SomeException _) -> return ()
conn <- socketConnection s
return (conn, sa)
closeListenSocket = sClose socket
-- | The connection setup action would be expensive. A good example
-- is initialization of TLS.
-- So, this converts the connection setup action to the connection maker
-- which will be executed after forking a new worker thread.
-- Then this calls 'runSettingsConnectionMaker' with the connection maker.
-- This allows the expensive computations to be performed
-- in a separate worker thread instead of the main server loop.
--
-- Since 1.3.5
runSettingsConnection :: Settings -> IO (Connection, SockAddr) -> Application -> IO ()
runSettingsConnection set getConn app = runSettingsConnectionMaker set getConnMaker app
where
getConnMaker = do
(conn, sa) <- getConn
return (return conn, sa)
-- | This modifies the connection maker so that it returns 'TCP' for 'Transport'
-- (i.e. plain HTTP) then calls 'runSettingsConnectionMakerSecure'.
runSettingsConnectionMaker :: Settings -> IO (IO Connection, SockAddr) -> Application -> IO ()
runSettingsConnectionMaker x y =
runSettingsConnectionMakerSecure x (toTCP <$> y)
where
toTCP = first ((, TCP) <$>)
----------------------------------------------------------------
-- | The core run function which takes 'Settings',
-- a connection maker and 'Application'.
-- The connection maker can return a connection of either plain HTTP
-- or HTTP over TLS.
--
-- Since 2.1.4
runSettingsConnectionMakerSecure :: Settings -> IO (IO (Connection, Transport), SockAddr) -> Application -> IO ()
runSettingsConnectionMakerSecure set getConnMaker app = do
settingsBeforeMainLoop set
counter <- newCounter
withII0 $ acceptConnection set getConnMaker app counter
where
withII0 action =
withTimeoutManager $ \tm ->
D.withDateCache $ \dc ->
F.withFdCache fdCacheDurationInSeconds $ \fdc ->
I.withFileInfoCache fdFileInfoDurationInSeconds $ \fic -> do
let ii0 = InternalInfo0 tm dc fdc fic
action ii0
!fdCacheDurationInSeconds = settingsFdCacheDuration set * 1000000
!fdFileInfoDurationInSeconds = settingsFileInfoCacheDuration set * 1000000
!timeoutInSeconds = settingsTimeout set * 1000000
withTimeoutManager f = case settingsManager set of
Just tm -> f tm
Nothing -> bracket
(T.initialize timeoutInSeconds)
T.stopManager
f
-- Note that there is a thorough discussion of the exception safety of the
-- following code at: https://github.com/yesodweb/wai/issues/146
--
-- We need to make sure of two things:
--
-- 1. Asynchronous exceptions are not blocked entirely in the main loop.
-- Doing so would make it impossible to kill the Warp thread.
--
-- 2. Once a connection maker is received via acceptNewConnection, the
-- connection is guaranteed to be closed, even in the presence of
-- async exceptions.
--
-- Our approach is explained in the comments below.
acceptConnection :: Settings
-> IO (IO (Connection, Transport), SockAddr)
-> Application
-> Counter
-> InternalInfo0
-> IO ()
acceptConnection set getConnMaker app counter ii0 = do
-- First mask all exceptions in acceptLoop. This is necessary to
-- ensure that no async exception is throw between the call to
-- acceptNewConnection and the registering of connClose.
void $ mask_ acceptLoop
gracefulShutdown counter
where
acceptLoop = do
-- Allow async exceptions before receiving the next connection maker.
allowInterrupt
-- acceptNewConnection will try to receive the next incoming
-- request. It returns a /connection maker/, not a connection,
-- since in some circumstances creating a working connection
-- from a raw socket may be an expensive operation, and this
-- expensive work should not be performed in the main event
-- loop. An example of something expensive would be TLS
-- negotiation.
mx <- acceptNewConnection
case mx of
Nothing -> return ()
Just (mkConn, addr) -> do
fork set mkConn addr app counter ii0
acceptLoop
acceptNewConnection = do
ex <- try getConnMaker
case ex of
Right x -> return $ Just x
Left e -> do
settingsOnException set Nothing $ toException e
if isFullErrorType (ioeGetErrorType e) then do
-- "resource exhausted (Too many open files)" may
-- happen by accept(). Wait a second hoping that
-- resource will be available.
threadDelay 1000000
acceptNewConnection
else
-- Assuming the listen socket is closed.
return Nothing
-- Fork a new worker thread for this connection maker, and ask for a
-- function to unmask (i.e., allow async exceptions to be thrown).
fork :: Settings
-> IO (Connection, Transport)
-> SockAddr
-> Application
-> Counter
-> InternalInfo0
-> IO ()
fork set mkConn addr app counter ii0 = settingsFork set $ \ unmask ->
-- Run the connection maker to get a new connection, and ensure
-- that the connection is closed. If the mkConn call throws an
-- exception, we will leak the connection. If the mkConn call is
-- vulnerable to attacks (e.g., Slowloris), we do nothing to
-- protect the server. It is therefore vital that mkConn is well
-- vetted.
--
-- We grab the connection before registering timeouts since the
-- timeouts will be useless during connection creation, due to the
-- fact that async exceptions are still masked.
bracket mkConn closeConn $ \(conn, transport) ->
-- We need to register a timeout handler for this thread, and
-- cancel that handler as soon as we exit.
bracket (T.registerKillThread (timeoutManager0 ii0)) T.cancel $ \th ->
let ii1 = toInternalInfo1 ii0 th
-- We now have fully registered a connection close handler
-- in the case of all exceptions, so it is safe to one
-- again allow async exceptions.
in unmask .
-- Call the user-supplied on exception code if any
-- exceptions are thrown.
handle (settingsOnException set Nothing) .
-- Call the user-supplied code for connection open and close events
bracket (onOpen addr) (onClose addr) $ \goingon ->
-- Actually serve this connection.
-- bracket with closeConn above ensures the connection is closed.
when goingon $ serveConnection conn ii1 addr transport set app
where
closeConn (conn, _transport) = connClose conn
onOpen adr = increase counter >> settingsOnOpen set adr
onClose adr _ = decrease counter >> settingsOnClose set adr
serveConnection :: Connection
-> InternalInfo1
-> SockAddr
-> Transport
-> Settings
-> Application
-> IO ()
serveConnection conn ii1 origAddr transport settings app = do
-- fixme: Upgrading to HTTP/2 should be supported.
(h2,bs) <- if isHTTP2 transport then
return (True, "")
else do
bs0 <- connRecv conn
if S.length bs0 >= 4 && "PRI " `S.isPrefixOf` bs0 then
return (True, bs0)
else
return (False, bs0)
if settingsHTTP2Enabled settings && h2 then do
recvN <- makeReceiveN bs (connRecv conn) (connRecvBuf conn)
-- fixme: origAddr
http2 conn ii1 origAddr transport settings recvN app
else do
istatus <- newIORef False
src <- mkSource (wrappedRecv conn th istatus (settingsSlowlorisSize settings))
writeIORef istatus True
leftoverSource src bs
addr <- getProxyProtocolAddr src
http1 addr istatus src `E.catch` \e -> do
sendErrorResponse addr istatus e
throwIO (e :: SomeException)
where
getProxyProtocolAddr src =
case settingsProxyProtocol settings of
ProxyProtocolNone ->
return origAddr
ProxyProtocolRequired -> do
seg <- readSource src
parseProxyProtocolHeader src seg
ProxyProtocolOptional -> do
seg <- readSource src
if S.isPrefixOf "PROXY " seg
then parseProxyProtocolHeader src seg
else do leftoverSource src seg
return origAddr
parseProxyProtocolHeader src seg = do
let (header,seg') = S.break (== 0x0d) seg -- 0x0d == CR
maybeAddr = case S.split 0x20 header of -- 0x20 == space
["PROXY","TCP4",clientAddr,_,clientPort,_] ->
case [x | (x, t) <- reads (decodeAscii clientAddr), null t] of
[a] -> Just (SockAddrInet (readInt clientPort)
(toHostAddress a))
_ -> Nothing
["PROXY","TCP6",clientAddr,_,clientPort,_] ->
case [x | (x, t) <- reads (decodeAscii clientAddr), null t] of
[a] -> Just (SockAddrInet6 (readInt clientPort)
0
(toHostAddress6 a)
0)
_ -> Nothing
("PROXY":"UNKNOWN":_) ->
Just origAddr
_ ->
Nothing
case maybeAddr of
Nothing -> throwIO (BadProxyHeader (decodeAscii header))
Just a -> do leftoverSource src (S.drop 2 seg') -- drop CRLF
return a
decodeAscii = map (chr . fromEnum) . S.unpack
th = threadHandle1 ii1
shouldSendErrorResponse se
| Just ConnectionClosedByPeer <- fromException se = False
| otherwise = True
sendErrorResponse addr istatus e = do
status <- readIORef istatus
when (shouldSendErrorResponse e && status) $ do
let ii = toInternalInfo ii1 0 -- dummy
dreq = dummyreq addr
void $ sendResponse settings conn ii dreq defaultIndexRequestHeader (return S.empty) (errorResponse e)
dummyreq addr = defaultRequest { remoteHost = addr }
errorResponse e = settingsOnExceptionResponse settings e
http1 addr istatus src = do
(req', mremainingRef, idxhdr, nextBodyFlush, ii) <- recvRequest settings conn ii1 addr src
let req = req' { isSecure = isTransportSecure transport }
keepAlive <- processRequest istatus src req mremainingRef idxhdr nextBodyFlush ii
`E.catch` \e -> do
-- Call the user-supplied exception handlers, passing the request.
sendErrorResponse addr istatus e
settingsOnException settings (Just req) e
-- Don't throw the error again to prevent calling settingsOnException twice.
return False
when keepAlive $ http1 addr istatus src
processRequest istatus src req mremainingRef idxhdr nextBodyFlush ii = do
-- Let the application run for as long as it wants
T.pause th
-- In the event that some scarce resource was acquired during
-- creating the request, we need to make sure that we don't get
-- an async exception before calling the ResponseSource.
keepAliveRef <- newIORef $ error "keepAliveRef not filled"
_ <- app req $ \res -> do
T.resume th
-- FIXME consider forcing evaluation of the res here to
-- send more meaningful error messages to the user.
-- However, it may affect performance.
writeIORef istatus False
keepAlive <- sendResponse settings conn ii req idxhdr (readSource src) res
writeIORef keepAliveRef keepAlive
return ResponseReceived
keepAlive <- readIORef keepAliveRef
-- We just send a Response and it takes a time to
-- receive a Request again. If we immediately call recv,
-- it is likely to fail and the IO manager works.
-- It is very costly. So, we yield to another Haskell
-- thread hoping that the next Request will arrive
-- when this Haskell thread will be re-scheduled.
-- This improves performance at least when
-- the number of cores is small.
Conc.yield
if not keepAlive then
return False
else
-- If there is an unknown or large amount of data to still be read
-- from the request body, simple drop this connection instead of
-- reading it all in to satisfy a keep-alive request.
case settingsMaximumBodyFlush settings of
Nothing -> do
flushEntireBody nextBodyFlush
T.resume th
return True
Just maxToRead -> do
let tryKeepAlive = do
-- flush the rest of the request body
isComplete <- flushBody nextBodyFlush maxToRead
if isComplete then do
T.resume th
return True
else
return False
case mremainingRef of
Just ref -> do
remaining <- readIORef ref
if remaining <= maxToRead then
tryKeepAlive
else
return False
Nothing -> tryKeepAlive
flushEntireBody :: IO ByteString -> IO ()
flushEntireBody src =
loop
where
loop = do
bs <- src
unless (S.null bs) loop
flushBody :: IO ByteString -- ^ get next chunk
-> Int -- ^ maximum to flush
-> IO Bool -- ^ True == flushed the entire body, False == we didn't
flushBody src =
loop
where
loop toRead = do
bs <- src
let toRead' = toRead - S.length bs
case () of
()
| S.null bs -> return True
| toRead' >= 0 -> loop toRead'
| otherwise -> return False
wrappedRecv :: Connection -> T.Handle -> IORef Bool -> Int -> IO ByteString
wrappedRecv Connection { connRecv = recv } th istatus slowlorisSize = do
bs <- recv
unless (S.null bs) $ do
writeIORef istatus True
when (S.length bs >= slowlorisSize) $ T.tickle th
return bs
-- Copied from: https://github.com/mzero/plush/blob/master/src/Plush/Server/Warp.hs
setSocketCloseOnExec :: Socket -> IO ()
#if WINDOWS
setSocketCloseOnExec _ = return ()
#else
setSocketCloseOnExec socket = F.setFileCloseOnExec $ fromIntegral $ fdSocket socket
#endif
gracefulShutdown :: Counter -> IO ()
gracefulShutdown counter = waitForZero counter
| erikd/wai | warp/Network/Wai/Handler/Warp/Run.hs | mit | 20,199 | 0 | 24 | 6,013 | 3,760 | 1,959 | 1,801 | 311 | 18 |
{-# LANGUAGE DataKinds #-}
{-# LANGUAGE DeriveGeneric #-}
{-# LANGUAGE FlexibleInstances #-}
{-# LANGUAGE GeneralizedNewtypeDeriving #-}
{-# LANGUAGE LambdaCase #-}
{-# LANGUAGE NoImplicitPrelude #-}
{-# LANGUAGE OverloadedStrings #-}
{-# LANGUAGE RecordWildCards #-}
{-# LANGUAGE TypeFamilies #-}
{-# OPTIONS_GHC -fno-warn-unused-imports #-}
-- Module : Network.AWS.Lambda.RemovePermission
-- Copyright : (c) 2013-2014 Brendan Hay <[email protected]>
-- License : This Source Code Form is subject to the terms of
-- the Mozilla Public License, v. 2.0.
-- A copy of the MPL can be found in the LICENSE file or
-- you can obtain it at http://mozilla.org/MPL/2.0/.
-- Maintainer : Brendan Hay <[email protected]>
-- Stability : experimental
-- Portability : non-portable (GHC extensions)
--
-- Derived from AWS service descriptions, licensed under Apache 2.0.
-- | You can remove individual permissions from an access policy associated with a
-- Lambda function by providing a Statement ID.
--
-- Note that removal of a permission will cause an active event source to lose
-- permission to the function.
--
-- You need permission for the 'lambda:RemovePermission' action.
--
-- <http://docs.aws.amazon.com/lambda/latest/dg/API_RemovePermission.html>
module Network.AWS.Lambda.RemovePermission
(
-- * Request
RemovePermission
-- ** Request constructor
, removePermission
-- ** Request lenses
, rpFunctionName
, rpStatementId
-- * Response
, RemovePermissionResponse
-- ** Response constructor
, removePermissionResponse
) where
import Network.AWS.Data (Object)
import Network.AWS.Prelude
import Network.AWS.Request.RestJSON
import Network.AWS.Lambda.Types
import qualified GHC.Exts
data RemovePermission = RemovePermission
{ _rpFunctionName :: Text
, _rpStatementId :: Text
} deriving (Eq, Ord, Read, Show)
-- | 'RemovePermission' constructor.
--
-- The fields accessible through corresponding lenses are:
--
-- * 'rpFunctionName' @::@ 'Text'
--
-- * 'rpStatementId' @::@ 'Text'
--
removePermission :: Text -- ^ 'rpFunctionName'
-> Text -- ^ 'rpStatementId'
-> RemovePermission
removePermission p1 p2 = RemovePermission
{ _rpFunctionName = p1
, _rpStatementId = p2
}
-- | Lambda function whose access policy you want to remove a permission from.
--
-- You can specify an unqualified function name (for example, "Thumbnail") or
-- you can specify Amazon Resource Name (ARN) of the function (for example,
-- "arn:aws:lambda:us-west-2:account-id:function:ThumbNail"). AWS Lambda also
-- allows you to specify only the account ID qualifier (for example,
-- "account-id:Thumbnail"). Note that the length constraint applies only to the
-- ARN. If you specify only the function name, it is limited to 64 character in
-- length.
rpFunctionName :: Lens' RemovePermission Text
rpFunctionName = lens _rpFunctionName (\s a -> s { _rpFunctionName = a })
-- | Statement ID of the permission to remove.
rpStatementId :: Lens' RemovePermission Text
rpStatementId = lens _rpStatementId (\s a -> s { _rpStatementId = a })
data RemovePermissionResponse = RemovePermissionResponse
deriving (Eq, Ord, Read, Show, Generic)
-- | 'RemovePermissionResponse' constructor.
removePermissionResponse :: RemovePermissionResponse
removePermissionResponse = RemovePermissionResponse
instance ToPath RemovePermission where
toPath RemovePermission{..} = mconcat
[ "/2015-03-31/functions/"
, toText _rpFunctionName
, "/versions/HEAD/policy/"
, toText _rpStatementId
]
instance ToQuery RemovePermission where
toQuery = const mempty
instance ToHeaders RemovePermission
instance ToJSON RemovePermission where
toJSON = const (toJSON Empty)
instance AWSRequest RemovePermission where
type Sv RemovePermission = Lambda
type Rs RemovePermission = RemovePermissionResponse
request = delete
response = nullResponse RemovePermissionResponse
| romanb/amazonka | amazonka-lambda/gen/Network/AWS/Lambda/RemovePermission.hs | mpl-2.0 | 4,172 | 0 | 9 | 870 | 441 | 273 | 168 | 57 | 1 |
-- Copyright (c) 1998-1999 Chris Okasaki.
-- See COPYRIGHT file for terms and conditions.
module RandList
{-# DEPRECATED "This module is unmaintained, and will disappear soon" #-}
(
-- type
Seq, -- instance of Sequence, Functor, Monad, MonadPlus
-- sequence operations
empty,single,cons,snoc,append,lview,lhead,ltail,rview,rhead,rtail,
null,size,concat,reverse,reverseOnto,fromList,toList,
map,concatMap,foldr,foldl,foldr1,foldl1,reducer,reducel,reduce1,
copy,tabulate,inBounds,lookup,lookupM,lookupWithDefault,update,adjust,
mapWithIndex,foldrWithIndex,foldlWithIndex,
take,drop,splitAt,subseq,filter,partition,takeWhile,dropWhile,splitWhile,
zip,zip3,zipWith,zipWith3,unzip,unzip3,unzipWith,unzipWith3,
-- documentation
moduleName,
-- re-export view type from EdisonPrelude for convenience
Maybe2(Just2,Nothing2)
) where
import Prelude hiding (concat,reverse,map,concatMap,foldr,foldl,foldr1,foldl1,
filter,takeWhile,dropWhile,lookup,take,drop,splitAt,
zip,zip3,zipWith,zipWith3,unzip,unzip3,null)
import EdisonPrelude(Maybe2(Just2,Nothing2))
import qualified Sequence as S( Sequence(..) )
import SequenceDefaults
import Monad
import QuickCheck
-- signatures for exported functions
moduleName :: String
empty :: Seq a
single :: a -> Seq a
cons :: a -> Seq a -> Seq a
snoc :: Seq a -> a -> Seq a
append :: Seq a -> Seq a -> Seq a
lview :: Seq a -> Maybe2 a (Seq a)
lhead :: Seq a -> a
ltail :: Seq a -> Seq a
rview :: Seq a -> Maybe2 (Seq a) a
rhead :: Seq a -> a
rtail :: Seq a -> Seq a
null :: Seq a -> Bool
size :: Seq a -> Int
concat :: Seq (Seq a) -> Seq a
reverse :: Seq a -> Seq a
reverseOnto :: Seq a -> Seq a -> Seq a
fromList :: [a] -> Seq a
toList :: Seq a -> [a]
map :: (a -> b) -> Seq a -> Seq b
concatMap :: (a -> Seq b) -> Seq a -> Seq b
foldr :: (a -> b -> b) -> b -> Seq a -> b
foldl :: (b -> a -> b) -> b -> Seq a -> b
foldr1 :: (a -> a -> a) -> Seq a -> a
foldl1 :: (a -> a -> a) -> Seq a -> a
reducer :: (a -> a -> a) -> a -> Seq a -> a
reducel :: (a -> a -> a) -> a -> Seq a -> a
reduce1 :: (a -> a -> a) -> Seq a -> a
copy :: Int -> a -> Seq a
tabulate :: Int -> (Int -> a) -> Seq a
inBounds :: Seq a -> Int -> Bool
lookup :: Seq a -> Int -> a
lookupM :: Seq a -> Int -> Maybe a
lookupWithDefault :: a -> Seq a -> Int -> a
update :: Int -> a -> Seq a -> Seq a
adjust :: (a -> a) -> Int -> Seq a -> Seq a
mapWithIndex :: (Int -> a -> b) -> Seq a -> Seq b
foldrWithIndex :: (Int -> a -> b -> b) -> b -> Seq a -> b
foldlWithIndex :: (b -> Int -> a -> b) -> b -> Seq a -> b
take :: Int -> Seq a -> Seq a
drop :: Int -> Seq a -> Seq a
splitAt :: Int -> Seq a -> (Seq a, Seq a)
subseq :: Int -> Int -> Seq a -> Seq a
filter :: (a -> Bool) -> Seq a -> Seq a
partition :: (a -> Bool) -> Seq a -> (Seq a, Seq a)
takeWhile :: (a -> Bool) -> Seq a -> Seq a
dropWhile :: (a -> Bool) -> Seq a -> Seq a
splitWhile :: (a -> Bool) -> Seq a -> (Seq a, Seq a)
zip :: Seq a -> Seq b -> Seq (a,b)
zip3 :: Seq a -> Seq b -> Seq c -> Seq (a,b,c)
zipWith :: (a -> b -> c) -> Seq a -> Seq b -> Seq c
zipWith3 :: (a -> b -> c -> d) -> Seq a -> Seq b -> Seq c -> Seq d
unzip :: Seq (a,b) -> (Seq a, Seq b)
unzip3 :: Seq (a,b,c) -> (Seq a, Seq b, Seq c)
unzipWith :: (a -> b) -> (a -> c) -> Seq a -> (Seq b, Seq c)
unzipWith3 :: (a -> b) -> (a -> c) -> (a -> d) -> Seq a -> (Seq b, Seq c, Seq d)
moduleName = "RandList"
-- Adapted from
-- Chris Okasaki. Purely Functional Data Structures. 1998.
-- Section 9.3.1.
-- and
-- Chris Okasaki. "Purely Functional Random Access Lists". FPCA'95,
-- pages 86-95.
data Tree a = L a | T a (Tree a) (Tree a) deriving (Eq)
data Seq a = E | C !Int (Tree a) (Seq a) --deriving (Eq)
-- want to derive Eq but can't because of GHC bug
half :: Int -> Int
half n = n `quot` 2 -- use a shift?
empty = E
single x = C 1 (L x) E
cons x xs@(C i s (C j t xs'))
| i == j = C (1 + i + j) (T x s t) xs'
cons x xs = C 1 (L x) xs
copy n x = if n <= 0 then E else buildTrees (1::Int) (L x)
where buildTrees j t
| j > n = takeTrees n (half j) (child t) E
| otherwise = buildTrees (1 + j + j) (T x t t)
takeTrees i j t xs
| i >= j = takeTrees (i - j) j t (C j t xs)
| i > 0 = takeTrees i (half j) (child t) xs
| otherwise = xs
child (T x s t) = t
lview E = Nothing2
lview (C _ (L x) xs) = Just2 x xs
lview (C i (T x s t) xs) = Just2 x (C j s (C j t xs))
where j = half i
lhead E = error "RandList.lhead: empty sequence"
lhead (C _ (L x) xs) = x
lhead (C _ (T x s t) xs) = x
ltail E = E
ltail (C _ (L x) xs) = xs
ltail (C i (T x s t) xs) = C j s (C j t xs)
where j = half i
rhead E = error "RandList.rhead: empty sequence"
rhead (C _ t E) = treeLast t
where treeLast (L x) = x
treeLast (T x s t) = treeLast t
rhead (C _ t xs) = rhead xs
null E = True
null _ = False
size xs = sz xs
where sz E = (0::Int)
sz (C j t xs) = j + sz xs
reverseOnto E ys = ys
reverseOnto (C _ t xs) ys = reverseOnto xs (revTree t ys)
where revTree (L x) ys = cons x ys
revTree (T x s t) ys = revTree t (revTree s (cons x ys))
map f E = E
map f (C j t xs) = C j (mapTree f t) (map f xs)
where mapTree f (L x) = L (f x)
mapTree f (T x s t) = T (f x) (mapTree f s) (mapTree f t)
foldr f e E = e
foldr f e (C _ t xs) = foldTree t (foldr f e xs)
where foldTree (L x) e = f x e
foldTree (T x s t) e = f x (foldTree s (foldTree t e))
foldl f e E = e
foldl f e (C _ t xs) = foldl f (foldTree e t) xs
where foldTree e (L x) = f e x
foldTree e (T x s t) = foldTree (foldTree (f e x) s) t
reduce1 f xs = case lview xs of
Nothing2 -> error "RandList.reduce1: empty seq"
Just2 x xs -> red1 x xs
where red1 x E = x
red1 x (C j t xs) = red1 (redTree x t) xs
redTree x (L y) = f x y
redTree x (T y s t) = redTree (redTree (f x y) s) t
inBounds xs i = inb xs i
where inb E i = False
inb (C j t xs) i
| i < j = (i >= 0)
| otherwise = inb xs (i - j)
lookup xs i = look xs i
where look E i = error "RandList.lookup: bad subscript"
look (C j t xs) i
| i < j = lookTree j t i
| otherwise = look xs (i - j)
lookTree _ (L x) i
| i == 0 = x
| otherwise = error "RandList.lookup: bad subscript"
lookTree j (T x s t) i
| i > k = lookTree k t (i - 1 - k)
| i /= 0 = lookTree k s (i - 1)
| otherwise = x
where k = half j
lookupM xs i = look xs i
where look E i = Nothing
look (C j t xs) i
| i < j = lookTree j t i
| otherwise = look xs (i - j)
lookTree _ (L x) i
| i == 0 = Just x
| otherwise = Nothing
lookTree j (T x s t) i
| i > k = lookTree k t (i - 1 - k)
| i /= 0 = lookTree k s (i - 1)
| otherwise = Just x
where k = half j
lookupWithDefault d xs i = look xs i
where look E i = d
look (C j t xs) i
| i < j = lookTree j t i
| otherwise = look xs (i - j)
lookTree _ (L x) i
| i == 0 = x
| otherwise = d
lookTree j (T x s t) i
| i > k = lookTree k t (i - 1 - k)
| i /= 0 = lookTree k s (i - 1)
| otherwise = x
where k = half j
update i y xs = upd i xs
where upd i E = E
upd i (C j t xs)
| i < j = C j (updTree i j t) xs
| otherwise = C j t (upd (i - j) xs)
updTree i j t@(L x)
| i == 0 = L y
| otherwise = t
updTree i j (T x s t)
| i > k = T x s (updTree (i - 1 - k) k t)
| i /= 0 = T x (updTree (i - 1) k s) t
| otherwise = T y s t
where k = half j
adjust f i xs = adj i xs
where adj i E = E
adj i (C j t xs)
| i < j = C j (adjTree i j t) xs
| otherwise = C j t (adj (i - j) xs)
adjTree i j t@(L x)
| i == 0 = L (f x)
| otherwise = t
adjTree i j (T x s t)
| i > k = T x s (adjTree (i - 1 - k) k t)
| i /= 0 = T x (adjTree (i - 1) k s) t
| otherwise = T (f x) s t
where k = half j
drop n xs = if n < 0 then xs else drp n xs
where drp i E = E
drp i (C j t xs)
| i < j = drpTree i j t xs
| otherwise = drp (i - j) xs
drpTree 0 j t xs = C j t xs
drpTree i j (L x) xs = error "RandList.drop: bug. Impossible case!"
drpTree i j (T x s t) xs
| i > k = drpTree (i - 1 - k) k t xs
| otherwise = drpTree (i - 1) k s (C k t xs)
where k = half j
-- the remaining functions all use defaults
snoc = snocUsingFoldr
append = appendUsingFoldr
rview = rviewDefault
rtail = rtailUsingLview
concat = concatUsingFoldr
reverse = reverseUsingReverseOnto
fromList = fromListUsingCons
toList = toListUsingFoldr
concatMap = concatMapUsingFoldr
foldr1 = foldr1UsingLview
foldl1 = foldl1UsingFoldl
reducer = reducerUsingReduce1
reducel = reducelUsingReduce1
tabulate = tabulateUsingLists
mapWithIndex = mapWithIndexUsingLists
foldrWithIndex = foldrWithIndexUsingLists
foldlWithIndex = foldlWithIndexUsingLists
take = takeUsingLists
splitAt = splitAtDefault
filter = filterUsingFoldr
partition = partitionUsingFoldr
subseq = subseqDefault
takeWhile = takeWhileUsingLview
dropWhile = dropWhileUsingLview
splitWhile = splitWhileUsingLview
-- for zips, could optimize by calculating which one is shorter and
-- retaining its shape
zip = zipUsingLists
zip3 = zip3UsingLists
zipWith = zipWithUsingLists
zipWith3 = zipWith3UsingLists
unzip = unzipUsingLists
unzip3 = unzip3UsingLists
unzipWith = unzipWithUsingLists
unzipWith3 = unzipWith3UsingLists
-- instances
instance S.Sequence Seq where
{empty = empty; single = single; cons = cons; snoc = snoc;
append = append; lview = lview; lhead = lhead; ltail = ltail;
rview = rview; rhead = rhead; rtail = rtail; null = null;
size = size; concat = concat; reverse = reverse;
reverseOnto = reverseOnto; fromList = fromList; toList = toList;
map = map; concatMap = concatMap; foldr = foldr; foldl = foldl;
foldr1 = foldr1; foldl1 = foldl1; reducer = reducer;
reducel = reducel; reduce1 = reduce1; copy = copy;
tabulate = tabulate; inBounds = inBounds; lookup = lookup;
lookupM = lookupM; lookupWithDefault = lookupWithDefault;
update = update; adjust = adjust; mapWithIndex = mapWithIndex;
foldrWithIndex = foldrWithIndex; foldlWithIndex = foldlWithIndex;
take = take; drop = drop; splitAt = splitAt; subseq = subseq;
filter = filter; partition = partition; takeWhile = takeWhile;
dropWhile = dropWhile; splitWhile = splitWhile; zip = zip;
zip3 = zip3; zipWith = zipWith; zipWith3 = zipWith3; unzip = unzip;
unzip3 = unzip3; unzipWith = unzipWith; unzipWith3 = unzipWith3;
instanceName s = moduleName}
instance Functor Seq where
fmap = map
instance Monad Seq where
return = single
xs >>= k = concatMap k xs
instance MonadPlus Seq where
mplus = append
mzero = empty
-- want to derive the following instance but can't because of GHC bug
instance Eq a => Eq (Seq a) where
C i tx xs == C j ty ys = (i == j) && (tx == ty) && (xs == ys)
E == E = True
_ == _ = False
instance Show a => Show (Seq a) where
show xs = show (toList xs)
instance Arbitrary a => Arbitrary (Seq a) where
arbitrary = do xs <- arbitrary
return (fromList xs)
coarbitrary xs = coarbitrary (toList xs)
| FranklinChen/hugs98-plus-Sep2006 | fptools/hslibs/data/edison/Seq/RandList.hs | bsd-3-clause | 12,098 | 2 | 13 | 4,006 | 5,495 | 2,831 | 2,664 | 295 | 5 |
{-#LANGUAGE NoImplicitPrelude #-}
{-#LANGUAGE LambdaCase #-}
{-#LANGUAGE ScopedTypeVariables #-}
{-#LANGUAGE OverloadedStrings #-}
module Web.Sprinkles.Cache.Memcached
where
import Web.Sprinkles.Prelude
import Web.Sprinkles.Cache
import Data.Time.Clock.POSIX
import qualified Data.HashMap.Strict as HashMap
import Data.Default
import qualified Database.Memcache.Client as Memcache
memcachedCache :: IO (Cache ByteString ByteString)
memcachedCache = do
let options :: Memcache.Options = def
withConnection =
bracket
(Memcache.newClient [] options)
(Memcache.quit)
expiry = 60
return
Cache
{ cacheGet = \key -> do
withConnection $ \client -> Memcache.gat client key expiry >>= \case
Just (val, _, _) -> return (Just val)
Nothing -> return Nothing
, cachePut = \key val -> do
withConnection $ \client ->
Memcache.set client key val 0 expiry
return ()
, cacheDelete = \key -> do
withConnection $ \client ->
Memcache.delete client key 0
return ()
, cacheVacuum = return 0
}
| tdammers/templar | src/Web/Sprinkles/Cache/Memcached.hs | bsd-3-clause | 1,249 | 0 | 21 | 409 | 300 | 164 | 136 | 34 | 2 |
{-
(c) The AQUA Project, Glasgow University, 1993-1998
\section[Simplify]{The main module of the simplifier}
-}
{-# LANGUAGE CPP #-}
module Simplify ( simplTopBinds, simplExpr, simplRules ) where
#include "HsVersions.h"
import DynFlags
import SimplMonad
import Type hiding ( substTy, substTyVar, extendTvSubst, extendCvSubst )
import SimplEnv
import SimplUtils
import FamInstEnv ( FamInstEnv )
import Literal ( litIsLifted ) --, mkMachInt ) -- temporalily commented out. See #8326
import Id
import MkId ( seqId, voidPrimId )
import MkCore ( mkImpossibleExpr, castBottomExpr )
import IdInfo
import Name ( Name, mkSystemVarName, isExternalName, getOccFS )
import Coercion hiding ( substCo, substCoVar )
import OptCoercion ( optCoercion )
import FamInstEnv ( topNormaliseType_maybe )
import DataCon ( DataCon, dataConWorkId, dataConRepStrictness
, isMarkedStrict, dataConRepArgTys ) --, dataConTyCon, dataConTag, fIRST_TAG )
--import TyCon ( isEnumerationTyCon ) -- temporalily commented out. See #8326
import CoreMonad ( Tick(..), SimplifierMode(..) )
import CoreSyn
import Demand ( StrictSig(..), dmdTypeDepth, isStrictDmd )
import PprCore ( pprCoreExpr )
import CoreUnfold
import CoreUtils
import CoreArity
--import PrimOp ( tagToEnumKey ) -- temporalily commented out. See #8326
import Rules ( mkRuleInfo, lookupRule, getRules )
import TysPrim ( voidPrimTy ) --, intPrimTy ) -- temporalily commented out. See #8326
import BasicTypes ( TopLevelFlag(..), isTopLevel, RecFlag(..) )
import MonadUtils ( foldlM, mapAccumLM, liftIO )
import Maybes ( orElse )
--import Unique ( hasKey ) -- temporalily commented out. See #8326
import Control.Monad
import Outputable
import FastString
import Pair
import Util
import ErrUtils
{-
The guts of the simplifier is in this module, but the driver loop for
the simplifier is in SimplCore.hs.
-----------------------------------------
*** IMPORTANT NOTE ***
-----------------------------------------
The simplifier used to guarantee that the output had no shadowing, but
it does not do so any more. (Actually, it never did!) The reason is
documented with simplifyArgs.
-----------------------------------------
*** IMPORTANT NOTE ***
-----------------------------------------
Many parts of the simplifier return a bunch of "floats" as well as an
expression. This is wrapped as a datatype SimplUtils.FloatsWith.
All "floats" are let-binds, not case-binds, but some non-rec lets may
be unlifted (with RHS ok-for-speculation).
-----------------------------------------
ORGANISATION OF FUNCTIONS
-----------------------------------------
simplTopBinds
- simplify all top-level binders
- for NonRec, call simplRecOrTopPair
- for Rec, call simplRecBind
------------------------------
simplExpr (applied lambda) ==> simplNonRecBind
simplExpr (Let (NonRec ...) ..) ==> simplNonRecBind
simplExpr (Let (Rec ...) ..) ==> simplify binders; simplRecBind
------------------------------
simplRecBind [binders already simplfied]
- use simplRecOrTopPair on each pair in turn
simplRecOrTopPair [binder already simplified]
Used for: recursive bindings (top level and nested)
top-level non-recursive bindings
Returns:
- check for PreInlineUnconditionally
- simplLazyBind
simplNonRecBind
Used for: non-top-level non-recursive bindings
beta reductions (which amount to the same thing)
Because it can deal with strict arts, it takes a
"thing-inside" and returns an expression
- check for PreInlineUnconditionally
- simplify binder, including its IdInfo
- if strict binding
simplStrictArg
mkAtomicArgs
completeNonRecX
else
simplLazyBind
addFloats
simplNonRecX: [given a *simplified* RHS, but an *unsimplified* binder]
Used for: binding case-binder and constr args in a known-constructor case
- check for PreInLineUnconditionally
- simplify binder
- completeNonRecX
------------------------------
simplLazyBind: [binder already simplified, RHS not]
Used for: recursive bindings (top level and nested)
top-level non-recursive bindings
non-top-level, but *lazy* non-recursive bindings
[must not be strict or unboxed]
Returns floats + an augmented environment, not an expression
- substituteIdInfo and add result to in-scope
[so that rules are available in rec rhs]
- simplify rhs
- mkAtomicArgs
- float if exposes constructor or PAP
- completeBind
completeNonRecX: [binder and rhs both simplified]
- if the the thing needs case binding (unlifted and not ok-for-spec)
build a Case
else
completeBind
addFloats
completeBind: [given a simplified RHS]
[used for both rec and non-rec bindings, top level and not]
- try PostInlineUnconditionally
- add unfolding [this is the only place we add an unfolding]
- add arity
Right hand sides and arguments
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
In many ways we want to treat
(a) the right hand side of a let(rec), and
(b) a function argument
in the same way. But not always! In particular, we would
like to leave these arguments exactly as they are, so they
will match a RULE more easily.
f (g x, h x)
g (+ x)
It's harder to make the rule match if we ANF-ise the constructor,
or eta-expand the PAP:
f (let { a = g x; b = h x } in (a,b))
g (\y. + x y)
On the other hand if we see the let-defns
p = (g x, h x)
q = + x
then we *do* want to ANF-ise and eta-expand, so that p and q
can be safely inlined.
Even floating lets out is a bit dubious. For let RHS's we float lets
out if that exposes a value, so that the value can be inlined more vigorously.
For example
r = let x = e in (x,x)
Here, if we float the let out we'll expose a nice constructor. We did experiments
that showed this to be a generally good thing. But it was a bad thing to float
lets out unconditionally, because that meant they got allocated more often.
For function arguments, there's less reason to expose a constructor (it won't
get inlined). Just possibly it might make a rule match, but I'm pretty skeptical.
So for the moment we don't float lets out of function arguments either.
Eta expansion
~~~~~~~~~~~~~~
For eta expansion, we want to catch things like
case e of (a,b) -> \x -> case a of (p,q) -> \y -> r
If the \x was on the RHS of a let, we'd eta expand to bring the two
lambdas together. And in general that's a good thing to do. Perhaps
we should eta expand wherever we find a (value) lambda? Then the eta
expansion at a let RHS can concentrate solely on the PAP case.
************************************************************************
* *
\subsection{Bindings}
* *
************************************************************************
-}
simplTopBinds :: SimplEnv -> [InBind] -> SimplM SimplEnv
simplTopBinds env0 binds0
= do { -- Put all the top-level binders into scope at the start
-- so that if a transformation rule has unexpectedly brought
-- anything into scope, then we don't get a complaint about that.
-- It's rather as if the top-level binders were imported.
-- See note [Glomming] in OccurAnal.
; env1 <- simplRecBndrs env0 (bindersOfBinds binds0)
; env2 <- simpl_binds env1 binds0
; freeTick SimplifierDone
; return env2 }
where
-- We need to track the zapped top-level binders, because
-- they should have their fragile IdInfo zapped (notably occurrence info)
-- That's why we run down binds and bndrs' simultaneously.
--
simpl_binds :: SimplEnv -> [InBind] -> SimplM SimplEnv
simpl_binds env [] = return env
simpl_binds env (bind:binds) = do { env' <- simpl_bind env bind
; simpl_binds env' binds }
simpl_bind env (Rec pairs) = simplRecBind env TopLevel pairs
simpl_bind env (NonRec b r) = do { (env', b') <- addBndrRules env b (lookupRecBndr env b)
; simplRecOrTopPair env' TopLevel NonRecursive b b' r }
{-
************************************************************************
* *
\subsection{Lazy bindings}
* *
************************************************************************
simplRecBind is used for
* recursive bindings only
-}
simplRecBind :: SimplEnv -> TopLevelFlag
-> [(InId, InExpr)]
-> SimplM SimplEnv
simplRecBind env0 top_lvl pairs0
= do { (env_with_info, triples) <- mapAccumLM add_rules env0 pairs0
; env1 <- go (zapFloats env_with_info) triples
; return (env0 `addRecFloats` env1) }
-- addFloats adds the floats from env1,
-- _and_ updates env0 with the in-scope set from env1
where
add_rules :: SimplEnv -> (InBndr,InExpr) -> SimplM (SimplEnv, (InBndr, OutBndr, InExpr))
-- Add the (substituted) rules to the binder
add_rules env (bndr, rhs)
= do { (env', bndr') <- addBndrRules env bndr (lookupRecBndr env bndr)
; return (env', (bndr, bndr', rhs)) }
go env [] = return env
go env ((old_bndr, new_bndr, rhs) : pairs)
= do { env' <- simplRecOrTopPair env top_lvl Recursive old_bndr new_bndr rhs
; go env' pairs }
{-
simplOrTopPair is used for
* recursive bindings (whether top level or not)
* top-level non-recursive bindings
It assumes the binder has already been simplified, but not its IdInfo.
-}
simplRecOrTopPair :: SimplEnv
-> TopLevelFlag -> RecFlag
-> InId -> OutBndr -> InExpr -- Binder and rhs
-> SimplM SimplEnv -- Returns an env that includes the binding
simplRecOrTopPair env top_lvl is_rec old_bndr new_bndr rhs
= do { dflags <- getDynFlags
; trace_bind dflags $
if preInlineUnconditionally dflags env top_lvl old_bndr rhs
-- Check for unconditional inline
then do tick (PreInlineUnconditionally old_bndr)
return (extendIdSubst env old_bndr (mkContEx env rhs))
else simplLazyBind env top_lvl is_rec old_bndr new_bndr rhs env }
where
trace_bind dflags thing_inside
| not (dopt Opt_D_verbose_core2core dflags)
= thing_inside
| otherwise
= pprTrace "SimplBind" (ppr old_bndr) thing_inside
-- trace_bind emits a trace for each top-level binding, which
-- helps to locate the tracing for inlining and rule firing
{-
simplLazyBind is used for
* [simplRecOrTopPair] recursive bindings (whether top level or not)
* [simplRecOrTopPair] top-level non-recursive bindings
* [simplNonRecE] non-top-level *lazy* non-recursive bindings
Nota bene:
1. It assumes that the binder is *already* simplified,
and is in scope, and its IdInfo too, except unfolding
2. It assumes that the binder type is lifted.
3. It does not check for pre-inline-unconditionally;
that should have been done already.
-}
simplLazyBind :: SimplEnv
-> TopLevelFlag -> RecFlag
-> InId -> OutId -- Binder, both pre-and post simpl
-- The OutId has IdInfo, except arity, unfolding
-> InExpr -> SimplEnv -- The RHS and its environment
-> SimplM SimplEnv
-- Precondition: rhs obeys the let/app invariant
simplLazyBind env top_lvl is_rec bndr bndr1 rhs rhs_se
= -- pprTrace "simplLazyBind" ((ppr bndr <+> ppr bndr1) $$ ppr rhs $$ ppr (seIdSubst rhs_se)) $
do { let rhs_env = rhs_se `setInScope` env
(tvs, body) = case collectTyAndValBinders rhs of
(tvs, [], body)
| surely_not_lam body -> (tvs, body)
_ -> ([], rhs)
surely_not_lam (Lam {}) = False
surely_not_lam (Tick t e)
| not (tickishFloatable t) = surely_not_lam e
-- eta-reduction could float
surely_not_lam _ = True
-- Do not do the "abstract tyyvar" thing if there's
-- a lambda inside, because it defeats eta-reduction
-- f = /\a. \x. g a x
-- should eta-reduce.
; (body_env, tvs') <- simplBinders rhs_env tvs
-- See Note [Floating and type abstraction] in SimplUtils
-- Simplify the RHS
; let rhs_cont = mkRhsStop (substTy body_env (exprType body))
; (body_env1, body1) <- simplExprF body_env body rhs_cont
-- ANF-ise a constructor or PAP rhs
; (body_env2, body2) <- prepareRhs top_lvl body_env1 bndr1 body1
; (env', rhs')
<- if not (doFloatFromRhs top_lvl is_rec False body2 body_env2)
then -- No floating, revert to body1
do { rhs' <- mkLam tvs' (wrapFloats body_env1 body1) rhs_cont
; return (env, rhs') }
else if null tvs then -- Simple floating
do { tick LetFloatFromLet
; return (addFloats env body_env2, body2) }
else -- Do type-abstraction first
do { tick LetFloatFromLet
; (poly_binds, body3) <- abstractFloats tvs' body_env2 body2
; rhs' <- mkLam tvs' body3 rhs_cont
; env' <- foldlM (addPolyBind top_lvl) env poly_binds
; return (env', rhs') }
; completeBind env' top_lvl bndr bndr1 rhs' }
{-
A specialised variant of simplNonRec used when the RHS is already simplified,
notably in knownCon. It uses case-binding where necessary.
-}
simplNonRecX :: SimplEnv
-> InId -- Old binder
-> OutExpr -- Simplified RHS
-> SimplM SimplEnv
-- Precondition: rhs satisfies the let/app invariant
simplNonRecX env bndr new_rhs
| isDeadBinder bndr -- Not uncommon; e.g. case (a,b) of c { (p,q) -> p }
= return env -- Here c is dead, and we avoid creating
-- the binding c = (a,b)
| Coercion co <- new_rhs
= return (extendCvSubst env bndr co)
| otherwise
= do { (env', bndr') <- simplBinder env bndr
; completeNonRecX NotTopLevel env' (isStrictId bndr) bndr bndr' new_rhs }
-- simplNonRecX is only used for NotTopLevel things
completeNonRecX :: TopLevelFlag -> SimplEnv
-> Bool
-> InId -- Old binder
-> OutId -- New binder
-> OutExpr -- Simplified RHS
-> SimplM SimplEnv
-- Precondition: rhs satisfies the let/app invariant
-- See Note [CoreSyn let/app invariant] in CoreSyn
completeNonRecX top_lvl env is_strict old_bndr new_bndr new_rhs
= do { (env1, rhs1) <- prepareRhs top_lvl (zapFloats env) new_bndr new_rhs
; (env2, rhs2) <-
if doFloatFromRhs NotTopLevel NonRecursive is_strict rhs1 env1
then do { tick LetFloatFromLet
; return (addFloats env env1, rhs1) } -- Add the floats to the main env
else return (env, wrapFloats env1 rhs1) -- Wrap the floats around the RHS
; completeBind env2 NotTopLevel old_bndr new_bndr rhs2 }
{-
{- No, no, no! Do not try preInlineUnconditionally in completeNonRecX
Doing so risks exponential behaviour, because new_rhs has been simplified once already
In the cases described by the folowing commment, postInlineUnconditionally will
catch many of the relevant cases.
-- This happens; for example, the case_bndr during case of
-- known constructor: case (a,b) of x { (p,q) -> ... }
-- Here x isn't mentioned in the RHS, so we don't want to
-- create the (dead) let-binding let x = (a,b) in ...
--
-- Similarly, single occurrences can be inlined vigourously
-- e.g. case (f x, g y) of (a,b) -> ....
-- If a,b occur once we can avoid constructing the let binding for them.
Furthermore in the case-binding case preInlineUnconditionally risks extra thunks
-- Consider case I# (quotInt# x y) of
-- I# v -> let w = J# v in ...
-- If we gaily inline (quotInt# x y) for v, we end up building an
-- extra thunk:
-- let w = J# (quotInt# x y) in ...
-- because quotInt# can fail.
| preInlineUnconditionally env NotTopLevel bndr new_rhs
= thing_inside (extendIdSubst env bndr (DoneEx new_rhs))
-}
----------------------------------
prepareRhs takes a putative RHS, checks whether it's a PAP or
constructor application and, if so, converts it to ANF, so that the
resulting thing can be inlined more easily. Thus
x = (f a, g b)
becomes
t1 = f a
t2 = g b
x = (t1,t2)
We also want to deal well cases like this
v = (f e1 `cast` co) e2
Here we want to make e1,e2 trivial and get
x1 = e1; x2 = e2; v = (f x1 `cast` co) v2
That's what the 'go' loop in prepareRhs does
-}
prepareRhs :: TopLevelFlag -> SimplEnv -> OutId -> OutExpr -> SimplM (SimplEnv, OutExpr)
-- Adds new floats to the env iff that allows us to return a good RHS
prepareRhs top_lvl env id (Cast rhs co) -- Note [Float coercions]
| Pair ty1 _ty2 <- coercionKind co -- Do *not* do this if rhs has an unlifted type
, not (isUnliftedType ty1) -- see Note [Float coercions (unlifted)]
= do { (env', rhs') <- makeTrivialWithInfo top_lvl env (getOccFS id) sanitised_info rhs
; return (env', Cast rhs' co) }
where
sanitised_info = vanillaIdInfo `setStrictnessInfo` strictnessInfo info
`setDemandInfo` demandInfo info
info = idInfo id
prepareRhs top_lvl env0 id rhs0
= do { (_is_exp, env1, rhs1) <- go 0 env0 rhs0
; return (env1, rhs1) }
where
go n_val_args env (Cast rhs co)
= do { (is_exp, env', rhs') <- go n_val_args env rhs
; return (is_exp, env', Cast rhs' co) }
go n_val_args env (App fun (Type ty))
= do { (is_exp, env', rhs') <- go n_val_args env fun
; return (is_exp, env', App rhs' (Type ty)) }
go n_val_args env (App fun arg)
= do { (is_exp, env', fun') <- go (n_val_args+1) env fun
; case is_exp of
True -> do { (env'', arg') <- makeTrivial top_lvl env' (getOccFS id) arg
; return (True, env'', App fun' arg') }
False -> return (False, env, App fun arg) }
go n_val_args env (Var fun)
= return (is_exp, env, Var fun)
where
is_exp = isExpandableApp fun n_val_args -- The fun a constructor or PAP
-- See Note [CONLIKE pragma] in BasicTypes
-- The definition of is_exp should match that in
-- OccurAnal.occAnalApp
go n_val_args env (Tick t rhs)
-- We want to be able to float bindings past this
-- tick. Non-scoping ticks don't care.
| tickishScoped t == NoScope
= do { (is_exp, env', rhs') <- go n_val_args env rhs
; return (is_exp, env', Tick t rhs') }
-- On the other hand, for scoping ticks we need to be able to
-- copy them on the floats, which in turn is only allowed if
-- we can obtain non-counting ticks.
| not (tickishCounts t) || tickishCanSplit t
= do { (is_exp, env', rhs') <- go n_val_args (zapFloats env) rhs
; let tickIt (id, expr) = (id, mkTick (mkNoCount t) expr)
floats' = seFloats $ env `addFloats` mapFloats env' tickIt
; return (is_exp, env' { seFloats = floats' }, Tick t rhs') }
go _ env other
= return (False, env, other)
{-
Note [Float coercions]
~~~~~~~~~~~~~~~~~~~~~~
When we find the binding
x = e `cast` co
we'd like to transform it to
x' = e
x = x `cast` co -- A trivial binding
There's a chance that e will be a constructor application or function, or something
like that, so moving the coercion to the usage site may well cancel the coercions
and lead to further optimisation. Example:
data family T a :: *
data instance T Int = T Int
foo :: Int -> Int -> Int
foo m n = ...
where
x = T m
go 0 = 0
go n = case x of { T m -> go (n-m) }
-- This case should optimise
Note [Preserve strictness when floating coercions]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
In the Note [Float coercions] transformation, keep the strictness info.
Eg
f = e `cast` co -- f has strictness SSL
When we transform to
f' = e -- f' also has strictness SSL
f = f' `cast` co -- f still has strictness SSL
Its not wrong to drop it on the floor, but better to keep it.
Note [Float coercions (unlifted)]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
BUT don't do [Float coercions] if 'e' has an unlifted type.
This *can* happen:
foo :: Int = (error (# Int,Int #) "urk")
`cast` CoUnsafe (# Int,Int #) Int
If do the makeTrivial thing to the error call, we'll get
foo = case error (# Int,Int #) "urk" of v -> v `cast` ...
But 'v' isn't in scope!
These strange casts can happen as a result of case-of-case
bar = case (case x of { T -> (# 2,3 #); F -> error "urk" }) of
(# p,q #) -> p+q
-}
makeTrivialArg :: SimplEnv -> ArgSpec -> SimplM (SimplEnv, ArgSpec)
makeTrivialArg env (ValArg e) = do
{ (env', e') <- makeTrivial NotTopLevel env (fsLit "arg") e
; return (env', ValArg e') }
makeTrivialArg env arg = return (env, arg) -- CastBy, TyArg
makeTrivial :: TopLevelFlag -> SimplEnv
-> FastString -- ^ a "friendly name" to build the new binder from
-> OutExpr -> SimplM (SimplEnv, OutExpr)
-- Binds the expression to a variable, if it's not trivial, returning the variable
makeTrivial top_lvl env context expr =
makeTrivialWithInfo top_lvl env context vanillaIdInfo expr
makeTrivialWithInfo :: TopLevelFlag -> SimplEnv
-> FastString
-- ^ a "friendly name" to build the new binder from
-> IdInfo -> OutExpr -> SimplM (SimplEnv, OutExpr)
-- Propagate strictness and demand info to the new binder
-- Note [Preserve strictness when floating coercions]
-- Returned SimplEnv has same substitution as incoming one
makeTrivialWithInfo top_lvl env context info expr
| exprIsTrivial expr -- Already trivial
|| not (bindingOk top_lvl expr expr_ty) -- Cannot trivialise
-- See Note [Cannot trivialise]
= return (env, expr)
| otherwise -- See Note [Take care] below
= do { uniq <- getUniqueM
; let name = mkSystemVarName uniq context
var = mkLocalIdOrCoVarWithInfo name expr_ty info
; env' <- completeNonRecX top_lvl env False var var expr
; expr' <- simplVar env' var
; return (env', expr') }
-- The simplVar is needed becase we're constructing a new binding
-- a = rhs
-- And if rhs is of form (rhs1 |> co), then we might get
-- a1 = rhs1
-- a = a1 |> co
-- and now a's RHS is trivial and can be substituted out, and that
-- is what completeNonRecX will do
-- To put it another way, it's as if we'd simplified
-- let var = e in var
where
expr_ty = exprType expr
bindingOk :: TopLevelFlag -> CoreExpr -> Type -> Bool
-- True iff we can have a binding of this expression at this level
-- Precondition: the type is the type of the expression
bindingOk top_lvl _ expr_ty
| isTopLevel top_lvl = not (isUnliftedType expr_ty)
| otherwise = True
{-
Note [Cannot trivialise]
~~~~~~~~~~~~~~~~~~~~~~~~
Consider tih
f :: Int -> Addr#
foo :: Bar
foo = Bar (f 3)
Then we can't ANF-ise foo, even though we'd like to, because
we can't make a top-level binding for the Addr# (f 3). And if
so we don't want to turn it into
foo = let x = f 3 in Bar x
because we'll just end up inlining x back, and that makes the
simplifier loop. Better not to ANF-ise it at all.
A case in point is literal strings (a MachStr is not regarded as
trivial):
foo = Ptr "blob"#
We don't want to ANF-ise this.
************************************************************************
* *
\subsection{Completing a lazy binding}
* *
************************************************************************
completeBind
* deals only with Ids, not TyVars
* takes an already-simplified binder and RHS
* is used for both recursive and non-recursive bindings
* is used for both top-level and non-top-level bindings
It does the following:
- tries discarding a dead binding
- tries PostInlineUnconditionally
- add unfolding [this is the only place we add an unfolding]
- add arity
It does *not* attempt to do let-to-case. Why? Because it is used for
- top-level bindings (when let-to-case is impossible)
- many situations where the "rhs" is known to be a WHNF
(so let-to-case is inappropriate).
Nor does it do the atomic-argument thing
-}
completeBind :: SimplEnv
-> TopLevelFlag -- Flag stuck into unfolding
-> InId -- Old binder
-> OutId -> OutExpr -- New binder and RHS
-> SimplM SimplEnv
-- completeBind may choose to do its work
-- * by extending the substitution (e.g. let x = y in ...)
-- * or by adding to the floats in the envt
--
-- Precondition: rhs obeys the let/app invariant
completeBind env top_lvl old_bndr new_bndr new_rhs
| isCoVar old_bndr
= case new_rhs of
Coercion co -> return (extendCvSubst env old_bndr co)
_ -> return (addNonRec env new_bndr new_rhs)
| otherwise
= ASSERT( isId new_bndr )
do { let old_info = idInfo old_bndr
old_unf = unfoldingInfo old_info
occ_info = occInfo old_info
-- Do eta-expansion on the RHS of the binding
-- See Note [Eta-expanding at let bindings] in SimplUtils
; (new_arity, final_rhs) <- tryEtaExpandRhs env new_bndr new_rhs
-- Simplify the unfolding
; new_unfolding <- simplLetUnfolding env top_lvl old_bndr final_rhs old_unf
; dflags <- getDynFlags
; if postInlineUnconditionally dflags env top_lvl new_bndr occ_info
final_rhs new_unfolding
-- Inline and discard the binding
then do { tick (PostInlineUnconditionally old_bndr)
; return (extendIdSubst env old_bndr (DoneEx final_rhs)) }
-- Use the substitution to make quite, quite sure that the
-- substitution will happen, since we are going to discard the binding
else
do { let info1 = idInfo new_bndr `setArityInfo` new_arity
-- Unfolding info: Note [Setting the new unfolding]
info2 = info1 `setUnfoldingInfo` new_unfolding
-- Demand info: Note [Setting the demand info]
--
-- We also have to nuke demand info if for some reason
-- eta-expansion *reduces* the arity of the binding to less
-- than that of the strictness sig. This can happen: see Note [Arity decrease].
info3 | isEvaldUnfolding new_unfolding
|| (case strictnessInfo info2 of
StrictSig dmd_ty -> new_arity < dmdTypeDepth dmd_ty)
= zapDemandInfo info2 `orElse` info2
| otherwise
= info2
final_id = new_bndr `setIdInfo` info3
; -- pprTrace "Binding" (ppr final_id <+> ppr new_unfolding) $
return (addNonRec env final_id final_rhs) } }
-- The addNonRec adds it to the in-scope set too
------------------------------
addPolyBind :: TopLevelFlag -> SimplEnv -> OutBind -> SimplM SimplEnv
-- Add a new binding to the environment, complete with its unfolding
-- but *do not* do postInlineUnconditionally, because we have already
-- processed some of the scope of the binding
-- We still want the unfolding though. Consider
-- let
-- x = /\a. let y = ... in Just y
-- in body
-- Then we float the y-binding out (via abstractFloats and addPolyBind)
-- but 'x' may well then be inlined in 'body' in which case we'd like the
-- opportunity to inline 'y' too.
--
-- INVARIANT: the arity is correct on the incoming binders
addPolyBind top_lvl env (NonRec poly_id rhs)
= do { unfolding <- simplLetUnfolding env top_lvl poly_id rhs noUnfolding
-- Assumes that poly_id did not have an INLINE prag
-- which is perhaps wrong. ToDo: think about this
; let final_id = setIdInfo poly_id $
idInfo poly_id `setUnfoldingInfo` unfolding
; return (addNonRec env final_id rhs) }
addPolyBind _ env bind@(Rec _)
= return (extendFloats env bind)
-- Hack: letrecs are more awkward, so we extend "by steam"
-- without adding unfoldings etc. At worst this leads to
-- more simplifier iterations
{- Note [Arity decrease]
~~~~~~~~~~~~~~~~~~~~~~~~
Generally speaking the arity of a binding should not decrease. But it *can*
legitimately happen because of RULES. Eg
f = g Int
where g has arity 2, will have arity 2. But if there's a rewrite rule
g Int --> h
where h has arity 1, then f's arity will decrease. Here's a real-life example,
which is in the output of Specialise:
Rec {
$dm {Arity 2} = \d.\x. op d
{-# RULES forall d. $dm Int d = $s$dm #-}
dInt = MkD .... opInt ...
opInt {Arity 1} = $dm dInt
$s$dm {Arity 0} = \x. op dInt }
Here opInt has arity 1; but when we apply the rule its arity drops to 0.
That's why Specialise goes to a little trouble to pin the right arity
on specialised functions too.
Note [Setting the demand info]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
If the unfolding is a value, the demand info may
go pear-shaped, so we nuke it. Example:
let x = (a,b) in
case x of (p,q) -> h p q x
Here x is certainly demanded. But after we've nuked
the case, we'll get just
let x = (a,b) in h a b x
and now x is not demanded (I'm assuming h is lazy)
This really happens. Similarly
let f = \x -> e in ...f..f...
After inlining f at some of its call sites the original binding may
(for example) be no longer strictly demanded.
The solution here is a bit ad hoc...
************************************************************************
* *
\subsection[Simplify-simplExpr]{The main function: simplExpr}
* *
************************************************************************
The reason for this OutExprStuff stuff is that we want to float *after*
simplifying a RHS, not before. If we do so naively we get quadratic
behaviour as things float out.
To see why it's important to do it after, consider this (real) example:
let t = f x
in fst t
==>
let t = let a = e1
b = e2
in (a,b)
in fst t
==>
let a = e1
b = e2
t = (a,b)
in
a -- Can't inline a this round, cos it appears twice
==>
e1
Each of the ==> steps is a round of simplification. We'd save a
whole round if we float first. This can cascade. Consider
let f = g d
in \x -> ...f...
==>
let f = let d1 = ..d.. in \y -> e
in \x -> ...f...
==>
let d1 = ..d..
in \x -> ...(\y ->e)...
Only in this second round can the \y be applied, and it
might do the same again.
-}
simplExpr :: SimplEnv -> CoreExpr -> SimplM CoreExpr
simplExpr env expr = simplExprC env expr (mkBoringStop expr_out_ty)
where
expr_out_ty :: OutType
expr_out_ty = substTy env (exprType expr)
simplExprC :: SimplEnv -> CoreExpr -> SimplCont -> SimplM CoreExpr
-- Simplify an expression, given a continuation
simplExprC env expr cont
= -- pprTrace "simplExprC" (ppr expr $$ ppr cont {- $$ ppr (seIdSubst env) -} $$ ppr (seFloats env) ) $
do { (env', expr') <- simplExprF (zapFloats env) expr cont
; -- pprTrace "simplExprC ret" (ppr expr $$ ppr expr') $
-- pprTrace "simplExprC ret3" (ppr (seInScope env')) $
-- pprTrace "simplExprC ret4" (ppr (seFloats env')) $
return (wrapFloats env' expr') }
--------------------------------------------------
simplExprF :: SimplEnv -> InExpr -> SimplCont
-> SimplM (SimplEnv, OutExpr)
simplExprF env e cont
= {- pprTrace "simplExprF" (vcat
[ ppr e
, text "cont =" <+> ppr cont
, text "inscope =" <+> ppr (seInScope env)
, text "tvsubst =" <+> ppr (seTvSubst env)
, text "idsubst =" <+> ppr (seIdSubst env)
, text "cvsubst =" <+> ppr (seCvSubst env)
{- , ppr (seFloats env) -}
]) $ -}
simplExprF1 env e cont
simplExprF1 :: SimplEnv -> InExpr -> SimplCont
-> SimplM (SimplEnv, OutExpr)
simplExprF1 env (Var v) cont = simplIdF env v cont
simplExprF1 env (Lit lit) cont = rebuild env (Lit lit) cont
simplExprF1 env (Tick t expr) cont = simplTick env t expr cont
simplExprF1 env (Cast body co) cont = simplCast env body co cont
simplExprF1 env (Coercion co) cont = simplCoercionF env co cont
simplExprF1 env (Type ty) cont = ASSERT( contIsRhsOrArg cont )
rebuild env (Type (substTy env ty)) cont
simplExprF1 env (App fun arg) cont
= simplExprF env fun $
case arg of
Type ty -> ApplyToTy { sc_arg_ty = substTy env ty
, sc_hole_ty = substTy env (exprType fun)
, sc_cont = cont }
_ -> ApplyToVal { sc_arg = arg, sc_env = env
, sc_dup = NoDup, sc_cont = cont }
simplExprF1 env expr@(Lam {}) cont
= simplLam env zapped_bndrs body cont
-- The main issue here is under-saturated lambdas
-- (\x1. \x2. e) arg1
-- Here x1 might have "occurs-once" occ-info, because occ-info
-- is computed assuming that a group of lambdas is applied
-- all at once. If there are too few args, we must zap the
-- occ-info, UNLESS the remaining binders are one-shot
where
(bndrs, body) = collectBinders expr
zapped_bndrs | need_to_zap = map zap bndrs
| otherwise = bndrs
need_to_zap = any zappable_bndr (drop n_args bndrs)
n_args = countArgs cont
-- NB: countArgs counts all the args (incl type args)
-- and likewise drop counts all binders (incl type lambdas)
zappable_bndr b = isId b && not (isOneShotBndr b)
zap b | isTyVar b = b
| otherwise = zapLamIdInfo b
simplExprF1 env (Case scrut bndr _ alts) cont
= simplExprF env scrut (Select { sc_dup = NoDup, sc_bndr = bndr
, sc_alts = alts
, sc_env = env, sc_cont = cont })
simplExprF1 env (Let (Rec pairs) body) cont
= do { env' <- simplRecBndrs env (map fst pairs)
-- NB: bndrs' don't have unfoldings or rules
-- We add them as we go down
; env'' <- simplRecBind env' NotTopLevel pairs
; simplExprF env'' body cont }
simplExprF1 env (Let (NonRec bndr rhs) body) cont
= simplNonRecE env bndr (rhs, env) ([], body) cont
---------------------------------
simplType :: SimplEnv -> InType -> SimplM OutType
-- Kept monadic just so we can do the seqType
simplType env ty
= -- pprTrace "simplType" (ppr ty $$ ppr (seTvSubst env)) $
seqType new_ty `seq` return new_ty
where
new_ty = substTy env ty
---------------------------------
simplCoercionF :: SimplEnv -> InCoercion -> SimplCont
-> SimplM (SimplEnv, OutExpr)
simplCoercionF env co cont
= do { co' <- simplCoercion env co
; rebuild env (Coercion co') cont }
simplCoercion :: SimplEnv -> InCoercion -> SimplM OutCoercion
simplCoercion env co
= let opt_co = optCoercion (getTCvSubst env) co
in seqCo opt_co `seq` return opt_co
-----------------------------------
-- | Push a TickIt context outwards past applications and cases, as
-- long as this is a non-scoping tick, to let case and application
-- optimisations apply.
simplTick :: SimplEnv -> Tickish Id -> InExpr -> SimplCont
-> SimplM (SimplEnv, OutExpr)
simplTick env tickish expr cont
-- A scoped tick turns into a continuation, so that we can spot
-- (scc t (\x . e)) in simplLam and eliminate the scc. If we didn't do
-- it this way, then it would take two passes of the simplifier to
-- reduce ((scc t (\x . e)) e').
-- NB, don't do this with counting ticks, because if the expr is
-- bottom, then rebuildCall will discard the continuation.
-- XXX: we cannot do this, because the simplifier assumes that
-- the context can be pushed into a case with a single branch. e.g.
-- scc<f> case expensive of p -> e
-- becomes
-- case expensive of p -> scc<f> e
--
-- So I'm disabling this for now. It just means we will do more
-- simplifier iterations that necessary in some cases.
-- | tickishScoped tickish && not (tickishCounts tickish)
-- = simplExprF env expr (TickIt tickish cont)
-- For unscoped or soft-scoped ticks, we are allowed to float in new
-- cost, so we simply push the continuation inside the tick. This
-- has the effect of moving the tick to the outside of a case or
-- application context, allowing the normal case and application
-- optimisations to fire.
| tickish `tickishScopesLike` SoftScope
= do { (env', expr') <- simplExprF env expr cont
; return (env', mkTick tickish expr')
}
-- Push tick inside if the context looks like this will allow us to
-- do a case-of-case - see Note [case-of-scc-of-case]
| Select {} <- cont, Just expr' <- push_tick_inside
= simplExprF env expr' cont
-- We don't want to move the tick, but we might still want to allow
-- floats to pass through with appropriate wrapping (or not, see
-- wrap_floats below)
--- | not (tickishCounts tickish) || tickishCanSplit tickish
-- = wrap_floats
| otherwise
= no_floating_past_tick
where
-- Try to push tick inside a case, see Note [case-of-scc-of-case].
push_tick_inside =
case expr0 of
Case scrut bndr ty alts
-> Just $ Case (tickScrut scrut) bndr ty (map tickAlt alts)
_other -> Nothing
where (ticks, expr0) = stripTicksTop movable (Tick tickish expr)
movable t = not (tickishCounts t) ||
t `tickishScopesLike` NoScope ||
tickishCanSplit t
tickScrut e = foldr mkTick e ticks
-- Alternatives get annotated with all ticks that scope in some way,
-- but we don't want to count entries.
tickAlt (c,bs,e) = (c,bs, foldr mkTick e ts_scope)
ts_scope = map mkNoCount $
filter (not . (`tickishScopesLike` NoScope)) ticks
no_floating_past_tick =
do { let (inc,outc) = splitCont cont
; (env', expr') <- simplExprF (zapFloats env) expr inc
; let tickish' = simplTickish env tickish
; (env'', expr'') <- rebuild (zapFloats env')
(wrapFloats env' expr')
(TickIt tickish' outc)
; return (addFloats env env'', expr'')
}
-- Alternative version that wraps outgoing floats with the tick. This
-- results in ticks being duplicated, as we don't make any attempt to
-- eliminate the tick if we re-inline the binding (because the tick
-- semantics allows unrestricted inlining of HNFs), so I'm not doing
-- this any more. FloatOut will catch any real opportunities for
-- floating.
--
-- wrap_floats =
-- do { let (inc,outc) = splitCont cont
-- ; (env', expr') <- simplExprF (zapFloats env) expr inc
-- ; let tickish' = simplTickish env tickish
-- ; let wrap_float (b,rhs) = (zapIdStrictness (setIdArity b 0),
-- mkTick (mkNoCount tickish') rhs)
-- -- when wrapping a float with mkTick, we better zap the Id's
-- -- strictness info and arity, because it might be wrong now.
-- ; let env'' = addFloats env (mapFloats env' wrap_float)
-- ; rebuild env'' expr' (TickIt tickish' outc)
-- }
simplTickish env tickish
| Breakpoint n ids <- tickish
= Breakpoint n (map (getDoneId . substId env) ids)
| otherwise = tickish
-- Push type application and coercion inside a tick
splitCont :: SimplCont -> (SimplCont, SimplCont)
splitCont cont@(ApplyToTy { sc_cont = tail }) = (cont { sc_cont = inc }, outc)
where (inc,outc) = splitCont tail
splitCont (CastIt co c) = (CastIt co inc, outc)
where (inc,outc) = splitCont c
splitCont other = (mkBoringStop (contHoleType other), other)
getDoneId (DoneId id) = id
getDoneId (DoneEx e) = getIdFromTrivialExpr e -- Note [substTickish] in CoreSubst
getDoneId other = pprPanic "getDoneId" (ppr other)
-- Note [case-of-scc-of-case]
-- It's pretty important to be able to transform case-of-case when
-- there's an SCC in the way. For example, the following comes up
-- in nofib/real/compress/Encode.hs:
--
-- case scctick<code_string.r1>
-- case $wcode_string_r13s wild_XC w1_s137 w2_s138 l_aje
-- of _ { (# ww1_s13f, ww2_s13g, ww3_s13h #) ->
-- (ww1_s13f, ww2_s13g, ww3_s13h)
-- }
-- of _ { (ww_s12Y, ww1_s12Z, ww2_s130) ->
-- tick<code_string.f1>
-- (ww_s12Y,
-- ww1_s12Z,
-- PTTrees.PT
-- @ GHC.Types.Char @ GHC.Types.Int wild2_Xj ww2_s130 r_ajf)
-- }
--
-- We really want this case-of-case to fire, because then the 3-tuple
-- will go away (indeed, the CPR optimisation is relying on this
-- happening). But the scctick is in the way - we need to push it
-- inside to expose the case-of-case. So we perform this
-- transformation on the inner case:
--
-- scctick c (case e of { p1 -> e1; ...; pn -> en })
-- ==>
-- case (scctick c e) of { p1 -> scc c e1; ...; pn -> scc c en }
--
-- So we've moved a constant amount of work out of the scc to expose
-- the case. We only do this when the continuation is interesting: in
-- for now, it has to be another Case (maybe generalise this later).
{-
************************************************************************
* *
\subsection{The main rebuilder}
* *
************************************************************************
-}
rebuild :: SimplEnv -> OutExpr -> SimplCont -> SimplM (SimplEnv, OutExpr)
-- At this point the substitution in the SimplEnv should be irrelevant
-- only the in-scope set and floats should matter
rebuild env expr cont
= case cont of
Stop {} -> return (env, expr)
TickIt t cont -> rebuild env (mkTick t expr) cont
CastIt co cont -> rebuild env (mkCast expr co) cont
-- NB: mkCast implements the (Coercion co |> g) optimisation
Select { sc_bndr = bndr, sc_alts = alts, sc_env = se, sc_cont = cont }
-> rebuildCase (se `setFloats` env) expr bndr alts cont
StrictArg info _ cont -> rebuildCall env (info `addValArgTo` expr) cont
StrictBind b bs body se cont -> do { env' <- simplNonRecX (se `setFloats` env) b expr
-- expr satisfies let/app since it started life
-- in a call to simplNonRecE
; simplLam env' bs body cont }
ApplyToTy { sc_arg_ty = ty, sc_cont = cont}
-> rebuild env (App expr (Type ty)) cont
ApplyToVal { sc_arg = arg, sc_env = se, sc_dup = dup_flag, sc_cont = cont}
-- See Note [Avoid redundant simplification]
| isSimplified dup_flag -> rebuild env (App expr arg) cont
| otherwise -> do { arg' <- simplExpr (se `setInScope` env) arg
; rebuild env (App expr arg') cont }
{-
************************************************************************
* *
\subsection{Lambdas}
* *
************************************************************************
-}
simplCast :: SimplEnv -> InExpr -> Coercion -> SimplCont
-> SimplM (SimplEnv, OutExpr)
simplCast env body co0 cont0
= do { co1 <- simplCoercion env co0
; cont1 <- addCoerce co1 cont0
; simplExprF env body cont1 }
where
addCoerce co cont = add_coerce co (coercionKind co) cont
add_coerce _co (Pair s1 k1) cont -- co :: ty~ty
| s1 `eqType` k1 = return cont -- is a no-op
add_coerce co1 (Pair s1 _k2) (CastIt co2 cont)
| (Pair _l1 t1) <- coercionKind co2
-- e |> (g1 :: S1~L) |> (g2 :: L~T1)
-- ==>
-- e, if S1=T1
-- e |> (g1 . g2 :: S1~T1) otherwise
--
-- For example, in the initial form of a worker
-- we may find (coerce T (coerce S (\x.e))) y
-- and we'd like it to simplify to e[y/x] in one round
-- of simplification
, s1 `eqType` t1 = return cont -- The coerces cancel out
| otherwise = return (CastIt (mkTransCo co1 co2) cont)
add_coerce co (Pair s1s2 _t1t2) cont@(ApplyToTy { sc_arg_ty = arg_ty, sc_cont = tail })
-- (f |> g) ty ---> (f ty) |> (g @ ty)
-- This implements the PushT rule from the paper
| isForAllTy s1s2
= do { cont' <- addCoerce new_cast tail
; return (cont { sc_cont = cont' }) }
where
new_cast = mkInstCo co (mkNomReflCo arg_ty)
add_coerce co (Pair s1s2 t1t2) (ApplyToVal { sc_arg = arg, sc_env = arg_se
, sc_dup = dup, sc_cont = cont })
| isFunTy s1s2 -- This implements the Push rule from the paper
, isFunTy t1t2 -- Check t1t2 to ensure 'arg' is a value arg
-- (e |> (g :: s1s2 ~ t1->t2)) f
-- ===>
-- (e (f |> (arg g :: t1~s1))
-- |> (res g :: s2->t2)
--
-- t1t2 must be a function type, t1->t2, because it's applied
-- to something but s1s2 might conceivably not be
--
-- When we build the ApplyTo we can't mix the out-types
-- with the InExpr in the argument, so we simply substitute
-- to make it all consistent. It's a bit messy.
-- But it isn't a common case.
--
-- Example of use: Trac #995
= do { (dup', arg_se', arg') <- simplArg env dup arg_se arg
; cont' <- addCoerce co2 cont
; return (ApplyToVal { sc_arg = mkCast arg' (mkSymCo co1)
, sc_env = arg_se'
, sc_dup = dup'
, sc_cont = cont' }) }
where
-- we split coercion t1->t2 ~ s1->s2 into t1 ~ s1 and
-- t2 ~ s2 with left and right on the curried form:
-- (->) t1 t2 ~ (->) s1 s2
[co1, co2] = decomposeCo 2 co
add_coerce co _ cont = return (CastIt co cont)
simplArg :: SimplEnv -> DupFlag -> StaticEnv -> CoreExpr
-> SimplM (DupFlag, StaticEnv, OutExpr)
simplArg env dup_flag arg_env arg
| isSimplified dup_flag
= return (dup_flag, arg_env, arg)
| otherwise
= do { arg' <- simplExpr (arg_env `setInScope` env) arg
; return (Simplified, zapSubstEnv arg_env, arg') }
{-
************************************************************************
* *
\subsection{Lambdas}
* *
************************************************************************
Note [Zap unfolding when beta-reducing]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Lambda-bound variables can have stable unfoldings, such as
$j = \x. \b{Unf=Just x}. e
See Note [Case binders and join points] below; the unfolding for lets
us optimise e better. However when we beta-reduce it we want to
revert to using the actual value, otherwise we can end up in the
stupid situation of
let x = blah in
let b{Unf=Just x} = y
in ...b...
Here it'd be far better to drop the unfolding and use the actual RHS.
-}
simplLam :: SimplEnv -> [InId] -> InExpr -> SimplCont
-> SimplM (SimplEnv, OutExpr)
simplLam env [] body cont = simplExprF env body cont
-- Beta reduction
simplLam env (bndr:bndrs) body (ApplyToTy { sc_arg_ty = arg_ty, sc_cont = cont })
= do { tick (BetaReduction bndr)
; simplLam (extendTvSubst env bndr arg_ty) bndrs body cont }
simplLam env (bndr:bndrs) body (ApplyToVal { sc_arg = arg, sc_env = arg_se
, sc_cont = cont })
= do { tick (BetaReduction bndr)
; simplNonRecE env' (zap_unfolding bndr) (arg, arg_se) (bndrs, body) cont }
where
env' | Coercion co <- arg
= extendCvSubst env bndr co
| otherwise
= env
zap_unfolding bndr -- See Note [Zap unfolding when beta-reducing]
| isId bndr, isStableUnfolding (realIdUnfolding bndr)
= setIdUnfolding bndr NoUnfolding
| otherwise = bndr
-- discard a non-counting tick on a lambda. This may change the
-- cost attribution slightly (moving the allocation of the
-- lambda elsewhere), but we don't care: optimisation changes
-- cost attribution all the time.
simplLam env bndrs body (TickIt tickish cont)
| not (tickishCounts tickish)
= simplLam env bndrs body cont
-- Not enough args, so there are real lambdas left to put in the result
simplLam env bndrs body cont
= do { (env', bndrs') <- simplLamBndrs env bndrs
; body' <- simplExpr env' body
; new_lam <- mkLam bndrs' body' cont
; rebuild env' new_lam cont }
simplLamBndrs :: SimplEnv -> [InBndr] -> SimplM (SimplEnv, [OutBndr])
simplLamBndrs env bndrs = mapAccumLM simplLamBndr env bndrs
-------------
simplLamBndr :: SimplEnv -> Var -> SimplM (SimplEnv, Var)
-- Used for lambda binders. These sometimes have unfoldings added by
-- the worker/wrapper pass that must be preserved, because they can't
-- be reconstructed from context. For example:
-- f x = case x of (a,b) -> fw a b x
-- fw a b x{=(a,b)} = ...
-- The "{=(a,b)}" is an unfolding we can't reconstruct otherwise.
simplLamBndr env bndr
| isId bndr && hasSomeUnfolding old_unf -- Special case
= do { (env1, bndr1) <- simplBinder env bndr
; unf' <- simplUnfolding env1 NotTopLevel bndr old_unf
; let bndr2 = bndr1 `setIdUnfolding` unf'
; return (modifyInScope env1 bndr2, bndr2) }
| otherwise
= simplBinder env bndr -- Normal case
where
old_unf = idUnfolding bndr
------------------
simplNonRecE :: SimplEnv
-> InBndr -- The binder
-> (InExpr, SimplEnv) -- Rhs of binding (or arg of lambda)
-> ([InBndr], InExpr) -- Body of the let/lambda
-- \xs.e
-> SimplCont
-> SimplM (SimplEnv, OutExpr)
-- simplNonRecE is used for
-- * non-top-level non-recursive lets in expressions
-- * beta reduction
--
-- It deals with strict bindings, via the StrictBind continuation,
-- which may abort the whole process
--
-- Precondition: rhs satisfies the let/app invariant
-- Note [CoreSyn let/app invariant] in CoreSyn
--
-- The "body" of the binding comes as a pair of ([InId],InExpr)
-- representing a lambda; so we recurse back to simplLam
-- Why? Because of the binder-occ-info-zapping done before
-- the call to simplLam in simplExprF (Lam ...)
-- First deal with type applications and type lets
-- (/\a. e) (Type ty) and (let a = Type ty in e)
simplNonRecE env bndr (Type ty_arg, rhs_se) (bndrs, body) cont
= ASSERT( isTyVar bndr )
do { ty_arg' <- simplType (rhs_se `setInScope` env) ty_arg
; simplLam (extendTvSubst env bndr ty_arg') bndrs body cont }
simplNonRecE env bndr (rhs, rhs_se) (bndrs, body) cont
= do dflags <- getDynFlags
case () of
_ | preInlineUnconditionally dflags env NotTopLevel bndr rhs
-> do { tick (PreInlineUnconditionally bndr)
; -- pprTrace "preInlineUncond" (ppr bndr <+> ppr rhs) $
simplLam (extendIdSubst env bndr (mkContEx rhs_se rhs)) bndrs body cont }
| isStrictId bndr -- Includes coercions
-> simplExprF (rhs_se `setFloats` env) rhs
(StrictBind bndr bndrs body env cont)
| otherwise
-> ASSERT( not (isTyVar bndr) )
do { (env1, bndr1) <- simplNonRecBndr env bndr
; (env2, bndr2) <- addBndrRules env1 bndr bndr1
; env3 <- simplLazyBind env2 NotTopLevel NonRecursive bndr bndr2 rhs rhs_se
; simplLam env3 bndrs body cont }
{-
************************************************************************
* *
Variables
* *
************************************************************************
-}
simplVar :: SimplEnv -> InVar -> SimplM OutExpr
-- Look up an InVar in the environment
simplVar env var
| isTyVar var = return (Type (substTyVar env var))
| isCoVar var = return (Coercion (substCoVar env var))
| otherwise
= case substId env var of
DoneId var1 -> return (Var var1)
DoneEx e -> return e
ContEx tvs cvs ids e -> simplExpr (setSubstEnv env tvs cvs ids) e
simplIdF :: SimplEnv -> InId -> SimplCont -> SimplM (SimplEnv, OutExpr)
simplIdF env var cont
= case substId env var of
DoneEx e -> simplExprF (zapSubstEnv env) e cont
ContEx tvs cvs ids e -> simplExprF (setSubstEnv env tvs cvs ids) e cont
DoneId var1 -> completeCall env var1 cont
-- Note [zapSubstEnv]
-- The template is already simplified, so don't re-substitute.
-- This is VITAL. Consider
-- let x = e in
-- let y = \z -> ...x... in
-- \ x -> ...y...
-- We'll clone the inner \x, adding x->x' in the id_subst
-- Then when we inline y, we must *not* replace x by x' in
-- the inlined copy!!
---------------------------------------------------------
-- Dealing with a call site
completeCall :: SimplEnv -> OutId -> SimplCont -> SimplM (SimplEnv, OutExpr)
completeCall env var cont
= do { ------------- Try inlining ----------------
dflags <- getDynFlags
; let (lone_variable, arg_infos, call_cont) = contArgs cont
n_val_args = length arg_infos
interesting_cont = interestingCallContext call_cont
unfolding = activeUnfolding env var
maybe_inline = callSiteInline dflags var unfolding
lone_variable arg_infos interesting_cont
; case maybe_inline of {
Just expr -- There is an inlining!
-> do { checkedTick (UnfoldingDone var)
; dump_inline dflags expr cont
; simplExprF (zapSubstEnv env) expr cont }
; Nothing -> do -- No inlining!
{ rule_base <- getSimplRules
; let info = mkArgInfo var (getRules rule_base var) n_val_args call_cont
; rebuildCall env info cont
}}}
where
dump_inline dflags unfolding cont
| not (dopt Opt_D_dump_inlinings dflags) = return ()
| not (dopt Opt_D_verbose_core2core dflags)
= when (isExternalName (idName var)) $
liftIO $ printOutputForUser dflags alwaysQualify $
sep [text "Inlining done:", nest 4 (ppr var)]
| otherwise
= liftIO $ printOutputForUser dflags alwaysQualify $
sep [text "Inlining done: " <> ppr var,
nest 4 (vcat [text "Inlined fn: " <+> nest 2 (ppr unfolding),
text "Cont: " <+> ppr cont])]
rebuildCall :: SimplEnv
-> ArgInfo
-> SimplCont
-> SimplM (SimplEnv, OutExpr)
rebuildCall env (ArgInfo { ai_fun = fun, ai_args = rev_args, ai_strs = [] }) cont
-- When we run out of strictness args, it means
-- that the call is definitely bottom; see SimplUtils.mkArgInfo
-- Then we want to discard the entire strict continuation. E.g.
-- * case (error "hello") of { ... }
-- * (error "Hello") arg
-- * f (error "Hello") where f is strict
-- etc
-- Then, especially in the first of these cases, we'd like to discard
-- the continuation, leaving just the bottoming expression. But the
-- type might not be right, so we may have to add a coerce.
| not (contIsTrivial cont) -- Only do this if there is a non-trivial
= return (env, castBottomExpr res cont_ty) -- contination to discard, else we do it
where -- again and again!
res = argInfoExpr fun rev_args
cont_ty = contResultType cont
rebuildCall env info (CastIt co cont)
= rebuildCall env (addCastTo info co) cont
rebuildCall env info (ApplyToTy { sc_arg_ty = arg_ty, sc_cont = cont })
= rebuildCall env (info `addTyArgTo` arg_ty) cont
rebuildCall env info@(ArgInfo { ai_encl = encl_rules, ai_type = fun_ty
, ai_strs = str:strs, ai_discs = disc:discs })
(ApplyToVal { sc_arg = arg, sc_env = arg_se
, sc_dup = dup_flag, sc_cont = cont })
| isSimplified dup_flag -- See Note [Avoid redundant simplification]
= rebuildCall env (addValArgTo info' arg) cont
| str -- Strict argument
= -- pprTrace "Strict Arg" (ppr arg $$ ppr (seIdSubst env) $$ ppr (seInScope env)) $
simplExprF (arg_se `setFloats` env) arg
(StrictArg info' cci cont)
-- Note [Shadowing]
| otherwise -- Lazy argument
-- DO NOT float anything outside, hence simplExprC
-- There is no benefit (unlike in a let-binding), and we'd
-- have to be very careful about bogus strictness through
-- floating a demanded let.
= do { arg' <- simplExprC (arg_se `setInScope` env) arg
(mkLazyArgStop (funArgTy fun_ty) cci)
; rebuildCall env (addValArgTo info' arg') cont }
where
info' = info { ai_strs = strs, ai_discs = discs }
cci | encl_rules = RuleArgCtxt
| disc > 0 = DiscArgCtxt -- Be keener here
| otherwise = BoringCtxt -- Nothing interesting
rebuildCall env (ArgInfo { ai_fun = fun, ai_args = rev_args, ai_rules = rules }) cont
| null rules
= rebuild env (argInfoExpr fun rev_args) cont -- No rules, common case
| otherwise
= do { -- We've accumulated a simplified call in <fun,rev_args>
-- so try rewrite rules; see Note [RULEs apply to simplified arguments]
-- See also Note [Rules for recursive functions]
; let env' = zapSubstEnv env -- See Note [zapSubstEnv];
-- and NB that 'rev_args' are all fully simplified
; mb_rule <- tryRules env' rules fun (reverse rev_args) cont
; case mb_rule of {
Just (rule_rhs, cont') -> simplExprF env' rule_rhs cont'
-- Rules don't match
; Nothing -> rebuild env (argInfoExpr fun rev_args) cont -- No rules
} }
{-
Note [RULES apply to simplified arguments]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
It's very desirable to try RULES once the arguments have been simplified, because
doing so ensures that rule cascades work in one pass. Consider
{-# RULES g (h x) = k x
f (k x) = x #-}
...f (g (h x))...
Then we want to rewrite (g (h x)) to (k x) and only then try f's rules. If
we match f's rules against the un-simplified RHS, it won't match. This
makes a particularly big difference when superclass selectors are involved:
op ($p1 ($p2 (df d)))
We want all this to unravel in one sweep.
Note [Avoid redundant simplification]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Because RULES apply to simplified arguments, there's a danger of repeatedly
simplifying already-simplified arguments. An important example is that of
(>>=) d e1 e2
Here e1, e2 are simplified before the rule is applied, but don't really
participate in the rule firing. So we mark them as Simplified to avoid
re-simplifying them.
Note [Shadowing]
~~~~~~~~~~~~~~~~
This part of the simplifier may break the no-shadowing invariant
Consider
f (...(\a -> e)...) (case y of (a,b) -> e')
where f is strict in its second arg
If we simplify the innermost one first we get (...(\a -> e)...)
Simplifying the second arg makes us float the case out, so we end up with
case y of (a,b) -> f (...(\a -> e)...) e'
So the output does not have the no-shadowing invariant. However, there is
no danger of getting name-capture, because when the first arg was simplified
we used an in-scope set that at least mentioned all the variables free in its
static environment, and that is enough.
We can't just do innermost first, or we'd end up with a dual problem:
case x of (a,b) -> f e (...(\a -> e')...)
I spent hours trying to recover the no-shadowing invariant, but I just could
not think of an elegant way to do it. The simplifier is already knee-deep in
continuations. We have to keep the right in-scope set around; AND we have
to get the effect that finding (error "foo") in a strict arg position will
discard the entire application and replace it with (error "foo"). Getting
all this at once is TOO HARD!
************************************************************************
* *
Rewrite rules
* *
************************************************************************
-}
tryRules :: SimplEnv -> [CoreRule]
-> Id -> [ArgSpec] -> SimplCont
-> SimplM (Maybe (CoreExpr, SimplCont))
-- The SimplEnv already has zapSubstEnv applied to it
tryRules env rules fn args call_cont
| null rules
= return Nothing
{- Disabled until we fix #8326
| fn `hasKey` tagToEnumKey -- See Note [Optimising tagToEnum#]
, [_type_arg, val_arg] <- args
, Select dup bndr ((_,[],rhs1) : rest_alts) se cont <- call_cont
, isDeadBinder bndr
= do { dflags <- getDynFlags
; let enum_to_tag :: CoreAlt -> CoreAlt
-- Takes K -> e into tagK# -> e
-- where tagK# is the tag of constructor K
enum_to_tag (DataAlt con, [], rhs)
= ASSERT( isEnumerationTyCon (dataConTyCon con) )
(LitAlt tag, [], rhs)
where
tag = mkMachInt dflags (toInteger (dataConTag con - fIRST_TAG))
enum_to_tag alt = pprPanic "tryRules: tagToEnum" (ppr alt)
new_alts = (DEFAULT, [], rhs1) : map enum_to_tag rest_alts
new_bndr = setIdType bndr intPrimTy
-- The binder is dead, but should have the right type
; return (Just (val_arg, Select dup new_bndr new_alts se cont)) }
-}
| otherwise
= do { dflags <- getDynFlags
; case lookupRule dflags (getUnfoldingInRuleMatch env) (activeRule env)
fn (argInfoAppArgs args) rules of {
Nothing ->
do { nodump dflags -- This ensures that an empty file is written
; return Nothing } ; -- No rule matches
Just (rule, rule_rhs) ->
do { checkedTick (RuleFired (ru_name rule))
; let cont' = pushSimplifiedArgs env
(drop (ruleArity rule) args)
call_cont
-- (ruleArity rule) says how many args the rule consumed
; dump dflags rule rule_rhs
; return (Just (rule_rhs, cont')) }}}
where
dump dflags rule rule_rhs
| dopt Opt_D_dump_rule_rewrites dflags
= log_rule dflags Opt_D_dump_rule_rewrites "Rule fired" $ vcat
[ text "Rule:" <+> ftext (ru_name rule)
, text "Before:" <+> hang (ppr fn) 2 (sep (map ppr args))
, text "After: " <+> pprCoreExpr rule_rhs
, text "Cont: " <+> ppr call_cont ]
| dopt Opt_D_dump_rule_firings dflags
= log_rule dflags Opt_D_dump_rule_firings "Rule fired:" $
ftext (ru_name rule)
| otherwise
= return ()
nodump dflags
| dopt Opt_D_dump_rule_rewrites dflags
= liftIO $ dumpSDoc dflags alwaysQualify Opt_D_dump_rule_rewrites "" empty
| dopt Opt_D_dump_rule_firings dflags
= liftIO $ dumpSDoc dflags alwaysQualify Opt_D_dump_rule_firings "" empty
| otherwise
= return ()
log_rule dflags flag hdr details
= liftIO . dumpSDoc dflags alwaysQualify flag "" $
sep [text hdr, nest 4 details]
{-
Note [Optimising tagToEnum#]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~
If we have an enumeration data type:
data Foo = A | B | C
Then we want to transform
case tagToEnum# x of ==> case x of
A -> e1 DEFAULT -> e1
B -> e2 1# -> e2
C -> e3 2# -> e3
thereby getting rid of the tagToEnum# altogether. If there was a DEFAULT
alternative we retain it (remember it comes first). If not the case must
be exhaustive, and we reflect that in the transformed version by adding
a DEFAULT. Otherwise Lint complains that the new case is not exhaustive.
See #8317.
Note [Rules for recursive functions]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
You might think that we shouldn't apply rules for a loop breaker:
doing so might give rise to an infinite loop, because a RULE is
rather like an extra equation for the function:
RULE: f (g x) y = x+y
Eqn: f a y = a-y
But it's too drastic to disable rules for loop breakers.
Even the foldr/build rule would be disabled, because foldr
is recursive, and hence a loop breaker:
foldr k z (build g) = g k z
So it's up to the programmer: rules can cause divergence
************************************************************************
* *
Rebuilding a case expression
* *
************************************************************************
Note [Case elimination]
~~~~~~~~~~~~~~~~~~~~~~~
The case-elimination transformation discards redundant case expressions.
Start with a simple situation:
case x# of ===> let y# = x# in e
y# -> e
(when x#, y# are of primitive type, of course). We can't (in general)
do this for algebraic cases, because we might turn bottom into
non-bottom!
The code in SimplUtils.prepareAlts has the effect of generalise this
idea to look for a case where we're scrutinising a variable, and we
know that only the default case can match. For example:
case x of
0# -> ...
DEFAULT -> ...(case x of
0# -> ...
DEFAULT -> ...) ...
Here the inner case is first trimmed to have only one alternative, the
DEFAULT, after which it's an instance of the previous case. This
really only shows up in eliminating error-checking code.
Note that SimplUtils.mkCase combines identical RHSs. So
case e of ===> case e of DEFAULT -> r
True -> r
False -> r
Now again the case may be elminated by the CaseElim transformation.
This includes things like (==# a# b#)::Bool so that we simplify
case ==# a# b# of { True -> x; False -> x }
to just
x
This particular example shows up in default methods for
comparison operations (e.g. in (>=) for Int.Int32)
Note [Case elimination: lifted case]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
If a case over a lifted type has a single alternative, and is being used
as a strict 'let' (all isDeadBinder bndrs), we may want to do this
transformation:
case e of r ===> let r = e in ...r...
_ -> ...r...
(a) 'e' is already evaluated (it may so if e is a variable)
Specifically we check (exprIsHNF e). In this case
we can just allocate the WHNF directly with a let.
or
(b) 'x' is not used at all and e is ok-for-speculation
The ok-for-spec bit checks that we don't lose any
exceptions or divergence.
NB: it'd be *sound* to switch from case to let if the
scrutinee was not yet WHNF but was guaranteed to
converge; but sticking with case means we won't build a
thunk
or
(c) 'x' is used strictly in the body, and 'e' is a variable
Then we can just substitute 'e' for 'x' in the body.
See Note [Eliminating redundant seqs]
For (b), the "not used at all" test is important. Consider
case (case a ># b of { True -> (p,q); False -> (q,p) }) of
r -> blah
The scrutinee is ok-for-speculation (it looks inside cases), but we do
not want to transform to
let r = case a ># b of { True -> (p,q); False -> (q,p) }
in blah
because that builds an unnecessary thunk.
Note [Eliminating redundant seqs]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
If we have this:
case x of r { _ -> ..r.. }
where 'r' is used strictly in (..r..), the case is effectively a 'seq'
on 'x', but since 'r' is used strictly anyway, we can safely transform to
(...x...)
Note that this can change the error behaviour. For example, we might
transform
case x of { _ -> error "bad" }
--> error "bad"
which is might be puzzling if 'x' currently lambda-bound, but later gets
let-bound to (error "good").
Nevertheless, the paper "A semantics for imprecise exceptions" allows
this transformation. If you want to fix the evaluation order, use
'pseq'. See Trac #8900 for an example where the loss of this
transformation bit us in practice.
See also Note [Empty case alternatives] in CoreSyn.
Just for reference, the original code (added Jan 13) looked like this:
|| case_bndr_evald_next rhs
case_bndr_evald_next :: CoreExpr -> Bool
-- See Note [Case binder next]
case_bndr_evald_next (Var v) = v == case_bndr
case_bndr_evald_next (Cast e _) = case_bndr_evald_next e
case_bndr_evald_next (App e _) = case_bndr_evald_next e
case_bndr_evald_next (Case e _ _ _) = case_bndr_evald_next e
case_bndr_evald_next _ = False
(This came up when fixing Trac #7542. See also Note [Eta reduction of
an eval'd function] in CoreUtils.)
Note [Case elimination: unlifted case]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Consider
case a +# b of r -> ...r...
Then we do case-elimination (to make a let) followed by inlining,
to get
.....(a +# b)....
If we have
case indexArray# a i of r -> ...r...
we might like to do the same, and inline the (indexArray# a i).
But indexArray# is not okForSpeculation, so we don't build a let
in rebuildCase (lest it get floated *out*), so the inlining doesn't
happen either.
This really isn't a big deal I think. The let can be
Further notes about case elimination
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Consider: test :: Integer -> IO ()
test = print
Turns out that this compiles to:
Print.test
= \ eta :: Integer
eta1 :: Void# ->
case PrelNum.< eta PrelNum.zeroInteger of wild { __DEFAULT ->
case hPutStr stdout
(PrelNum.jtos eta ($w[] @ Char))
eta1
of wild1 { (# new_s, a4 #) -> PrelIO.lvl23 new_s }}
Notice the strange '<' which has no effect at all. This is a funny one.
It started like this:
f x y = if x < 0 then jtos x
else if y==0 then "" else jtos x
At a particular call site we have (f v 1). So we inline to get
if v < 0 then jtos x
else if 1==0 then "" else jtos x
Now simplify the 1==0 conditional:
if v<0 then jtos v else jtos v
Now common-up the two branches of the case:
case (v<0) of DEFAULT -> jtos v
Why don't we drop the case? Because it's strict in v. It's technically
wrong to drop even unnecessary evaluations, and in practice they
may be a result of 'seq' so we *definitely* don't want to drop those.
I don't really know how to improve this situation.
-}
---------------------------------------------------------
-- Eliminate the case if possible
rebuildCase, reallyRebuildCase
:: SimplEnv
-> OutExpr -- Scrutinee
-> InId -- Case binder
-> [InAlt] -- Alternatives (inceasing order)
-> SimplCont
-> SimplM (SimplEnv, OutExpr)
--------------------------------------------------
-- 1. Eliminate the case if there's a known constructor
--------------------------------------------------
rebuildCase env scrut case_bndr alts cont
| Lit lit <- scrut -- No need for same treatment as constructors
-- because literals are inlined more vigorously
, not (litIsLifted lit)
= do { tick (KnownBranch case_bndr)
; case findAlt (LitAlt lit) alts of
Nothing -> missingAlt env case_bndr alts cont
Just (_, bs, rhs) -> simple_rhs bs rhs }
| Just (con, ty_args, other_args) <- exprIsConApp_maybe (getUnfoldingInRuleMatch env) scrut
-- Works when the scrutinee is a variable with a known unfolding
-- as well as when it's an explicit constructor application
= do { tick (KnownBranch case_bndr)
; case findAlt (DataAlt con) alts of
Nothing -> missingAlt env case_bndr alts cont
Just (DEFAULT, bs, rhs) -> simple_rhs bs rhs
Just (_, bs, rhs) -> knownCon env scrut con ty_args other_args
case_bndr bs rhs cont
}
where
simple_rhs bs rhs = ASSERT( null bs )
do { env' <- simplNonRecX env case_bndr scrut
-- scrut is a constructor application,
-- hence satisfies let/app invariant
; simplExprF env' rhs cont }
--------------------------------------------------
-- 2. Eliminate the case if scrutinee is evaluated
--------------------------------------------------
rebuildCase env scrut case_bndr alts@[(_, bndrs, rhs)] cont
-- See if we can get rid of the case altogether
-- See Note [Case elimination]
-- mkCase made sure that if all the alternatives are equal,
-- then there is now only one (DEFAULT) rhs
-- 2a. Dropping the case altogether, if
-- a) it binds nothing (so it's really just a 'seq')
-- b) evaluating the scrutinee has no side effects
| is_plain_seq
, exprOkForSideEffects scrut
-- The entire case is dead, so we can drop it
-- if the scrutinee converges without having imperative
-- side effects or raising a Haskell exception
-- See Note [PrimOp can_fail and has_side_effects] in PrimOp
= simplExprF env rhs cont
-- 2b. Turn the case into a let, if
-- a) it binds only the case-binder
-- b) unlifted case: the scrutinee is ok-for-speculation
-- lifted case: the scrutinee is in HNF (or will later be demanded)
| all_dead_bndrs
, if is_unlifted
then exprOkForSpeculation scrut -- See Note [Case elimination: unlifted case]
else exprIsHNF scrut -- See Note [Case elimination: lifted case]
|| scrut_is_demanded_var scrut
= do { tick (CaseElim case_bndr)
; env' <- simplNonRecX env case_bndr scrut
; simplExprF env' rhs cont }
-- 2c. Try the seq rules if
-- a) it binds only the case binder
-- b) a rule for seq applies
-- See Note [User-defined RULES for seq] in MkId
| is_plain_seq
= do { let scrut_ty = exprType scrut
rhs_ty = substTy env (exprType rhs)
out_args = [ TyArg { as_arg_ty = scrut_ty
, as_hole_ty = seq_id_ty }
, TyArg { as_arg_ty = rhs_ty
, as_hole_ty = piResultTy seq_id_ty scrut_ty }
, ValArg scrut]
rule_cont = ApplyToVal { sc_dup = NoDup, sc_arg = rhs
, sc_env = env, sc_cont = cont }
env' = zapSubstEnv env
-- Lazily evaluated, so we don't do most of this
; rule_base <- getSimplRules
; mb_rule <- tryRules env' (getRules rule_base seqId) seqId out_args rule_cont
; case mb_rule of
Just (rule_rhs, cont') -> simplExprF env' rule_rhs cont'
Nothing -> reallyRebuildCase env scrut case_bndr alts cont }
where
is_unlifted = isUnliftedType (idType case_bndr)
all_dead_bndrs = all isDeadBinder bndrs -- bndrs are [InId]
is_plain_seq = all_dead_bndrs && isDeadBinder case_bndr -- Evaluation *only* for effect
seq_id_ty = idType seqId
scrut_is_demanded_var :: CoreExpr -> Bool
-- See Note [Eliminating redundant seqs]
scrut_is_demanded_var (Cast s _) = scrut_is_demanded_var s
scrut_is_demanded_var (Var _) = isStrictDmd (idDemandInfo case_bndr)
scrut_is_demanded_var _ = False
rebuildCase env scrut case_bndr alts cont
= reallyRebuildCase env scrut case_bndr alts cont
--------------------------------------------------
-- 3. Catch-all case
--------------------------------------------------
reallyRebuildCase env scrut case_bndr alts cont
= do { -- Prepare the continuation;
-- The new subst_env is in place
(env', dup_cont, nodup_cont) <- prepareCaseCont env alts cont
-- Simplify the alternatives
; (scrut', case_bndr', alts') <- simplAlts env' scrut case_bndr alts dup_cont
; dflags <- getDynFlags
; let alts_ty' = contResultType dup_cont
; case_expr <- mkCase dflags scrut' case_bndr' alts_ty' alts'
-- Notice that rebuild gets the in-scope set from env', not alt_env
-- (which in any case is only build in simplAlts)
-- The case binder *not* scope over the whole returned case-expression
; rebuild env' case_expr nodup_cont }
{-
simplCaseBinder checks whether the scrutinee is a variable, v. If so,
try to eliminate uses of v in the RHSs in favour of case_bndr; that
way, there's a chance that v will now only be used once, and hence
inlined.
Historical note: we use to do the "case binder swap" in the Simplifier
so there were additional complications if the scrutinee was a variable.
Now the binder-swap stuff is done in the occurrence analyer; see
OccurAnal Note [Binder swap].
Note [knownCon occ info]
~~~~~~~~~~~~~~~~~~~~~~~~
If the case binder is not dead, then neither are the pattern bound
variables:
case <any> of x { (a,b) ->
case x of { (p,q) -> p } }
Here (a,b) both look dead, but come alive after the inner case is eliminated.
The point is that we bring into the envt a binding
let x = (a,b)
after the outer case, and that makes (a,b) alive. At least we do unless
the case binder is guaranteed dead.
Note [Case alternative occ info]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
When we are simply reconstructing a case (the common case), we always
zap the occurrence info on the binders in the alternatives. Even
if the case binder is dead, the scrutinee is usually a variable, and *that*
can bring the case-alternative binders back to life.
See Note [Add unfolding for scrutinee]
Note [Improving seq]
~~~~~~~~~~~~~~~~~~~
Consider
type family F :: * -> *
type instance F Int = Int
... case e of x { DEFAULT -> rhs } ...
where x::F Int. Then we'd like to rewrite (F Int) to Int, getting
case e `cast` co of x'::Int
I# x# -> let x = x' `cast` sym co
in rhs
so that 'rhs' can take advantage of the form of x'.
Notice that Note [Case of cast] (in OccurAnal) may then apply to the result.
Nota Bene: We only do the [Improving seq] transformation if the
case binder 'x' is actually used in the rhs; that is, if the case
is *not* a *pure* seq.
a) There is no point in adding the cast to a pure seq.
b) There is a good reason not to: doing so would interfere
with seq rules (Note [Built-in RULES for seq] in MkId).
In particular, this [Improving seq] thing *adds* a cast
while [Built-in RULES for seq] *removes* one, so they
just flip-flop.
You might worry about
case v of x { __DEFAULT ->
... case (v `cast` co) of y { I# -> ... }}
This is a pure seq (since x is unused), so [Improving seq] won't happen.
But it's ok: the simplifier will replace 'v' by 'x' in the rhs to get
case v of x { __DEFAULT ->
... case (x `cast` co) of y { I# -> ... }}
Now the outer case is not a pure seq, so [Improving seq] will happen,
and then the inner case will disappear.
The need for [Improving seq] showed up in Roman's experiments. Example:
foo :: F Int -> Int -> Int
foo t n = t `seq` bar n
where
bar 0 = 0
bar n = bar (n - case t of TI i -> i)
Here we'd like to avoid repeated evaluating t inside the loop, by
taking advantage of the `seq`.
At one point I did transformation in LiberateCase, but it's more
robust here. (Otherwise, there's a danger that we'll simply drop the
'seq' altogether, before LiberateCase gets to see it.)
-}
simplAlts :: SimplEnv
-> OutExpr
-> InId -- Case binder
-> [InAlt] -- Non-empty
-> SimplCont
-> SimplM (OutExpr, OutId, [OutAlt]) -- Includes the continuation
-- Like simplExpr, this just returns the simplified alternatives;
-- it does not return an environment
-- The returned alternatives can be empty, none are possible
simplAlts env scrut case_bndr alts cont'
= do { let env0 = zapFloats env
; (env1, case_bndr1) <- simplBinder env0 case_bndr
; fam_envs <- getFamEnvs
; (alt_env', scrut', case_bndr') <- improveSeq fam_envs env1 scrut
case_bndr case_bndr1 alts
; (imposs_deflt_cons, in_alts) <- prepareAlts scrut' case_bndr' alts
-- NB: it's possible that the returned in_alts is empty: this is handled
-- by the caller (rebuildCase) in the missingAlt function
; alts' <- mapM (simplAlt alt_env' (Just scrut') imposs_deflt_cons case_bndr' cont') in_alts
; -- pprTrace "simplAlts" (ppr case_bndr $$ ppr alts_ty $$ ppr alts_ty' $$ ppr alts $$ ppr cont') $
return (scrut', case_bndr', alts') }
------------------------------------
improveSeq :: (FamInstEnv, FamInstEnv) -> SimplEnv
-> OutExpr -> InId -> OutId -> [InAlt]
-> SimplM (SimplEnv, OutExpr, OutId)
-- Note [Improving seq]
improveSeq fam_envs env scrut case_bndr case_bndr1 [(DEFAULT,_,_)]
| not (isDeadBinder case_bndr) -- Not a pure seq! See Note [Improving seq]
, Just (co, ty2) <- topNormaliseType_maybe fam_envs (idType case_bndr1)
= do { case_bndr2 <- newId (fsLit "nt") ty2
; let rhs = DoneEx (Var case_bndr2 `Cast` mkSymCo co)
env2 = extendIdSubst env case_bndr rhs
; return (env2, scrut `Cast` co, case_bndr2) }
improveSeq _ env scrut _ case_bndr1 _
= return (env, scrut, case_bndr1)
------------------------------------
simplAlt :: SimplEnv
-> Maybe OutExpr -- The scrutinee
-> [AltCon] -- These constructors can't be present when
-- matching the DEFAULT alternative
-> OutId -- The case binder
-> SimplCont
-> InAlt
-> SimplM OutAlt
simplAlt env _ imposs_deflt_cons case_bndr' cont' (DEFAULT, bndrs, rhs)
= ASSERT( null bndrs )
do { let env' = addBinderUnfolding env case_bndr'
(mkOtherCon imposs_deflt_cons)
-- Record the constructors that the case-binder *can't* be.
; rhs' <- simplExprC env' rhs cont'
; return (DEFAULT, [], rhs') }
simplAlt env scrut' _ case_bndr' cont' (LitAlt lit, bndrs, rhs)
= ASSERT( null bndrs )
do { env' <- addAltUnfoldings env scrut' case_bndr' (Lit lit)
; rhs' <- simplExprC env' rhs cont'
; return (LitAlt lit, [], rhs') }
simplAlt env scrut' _ case_bndr' cont' (DataAlt con, vs, rhs)
= do { -- Deal with the pattern-bound variables
-- Mark the ones that are in ! positions in the
-- data constructor as certainly-evaluated.
-- NB: simplLamBinders preserves this eval info
; let vs_with_evals = add_evals (dataConRepStrictness con)
; (env', vs') <- simplLamBndrs env vs_with_evals
-- Bind the case-binder to (con args)
; let inst_tys' = tyConAppArgs (idType case_bndr')
con_app :: OutExpr
con_app = mkConApp2 con inst_tys' vs'
; env'' <- addAltUnfoldings env' scrut' case_bndr' con_app
; rhs' <- simplExprC env'' rhs cont'
; return (DataAlt con, vs', rhs') }
where
-- add_evals records the evaluated-ness of the bound variables of
-- a case pattern. This is *important*. Consider
-- data T = T !Int !Int
--
-- case x of { T a b -> T (a+1) b }
--
-- We really must record that b is already evaluated so that we don't
-- go and re-evaluate it when constructing the result.
-- See Note [Data-con worker strictness] in MkId.hs
add_evals the_strs
= go vs the_strs
where
go [] [] = []
go (v:vs') strs | isTyVar v = v : go vs' strs
go (v:vs') (str:strs)
| isMarkedStrict str = eval v : go vs' strs
| otherwise = zap v : go vs' strs
go _ _ = pprPanic "cat_evals"
(ppr con $$
ppr vs $$
ppr_with_length the_strs $$
ppr_with_length (dataConRepArgTys con) $$
ppr_with_length (dataConRepStrictness con))
where
ppr_with_length list
= ppr list <+> parens (text "length =" <+> ppr (length list))
-- NB: If this panic triggers, note that
-- NoStrictnessMark doesn't print!
zap v = zapIdOccInfo v -- See Note [Case alternative occ info]
eval v = zap v `setIdUnfolding` evaldUnfolding
addAltUnfoldings :: SimplEnv -> Maybe OutExpr -> OutId -> OutExpr -> SimplM SimplEnv
addAltUnfoldings env scrut case_bndr con_app
= do { dflags <- getDynFlags
; let con_app_unf = mkSimpleUnfolding dflags con_app
env1 = addBinderUnfolding env case_bndr con_app_unf
-- See Note [Add unfolding for scrutinee]
env2 = case scrut of
Just (Var v) -> addBinderUnfolding env1 v con_app_unf
Just (Cast (Var v) co) -> addBinderUnfolding env1 v $
mkSimpleUnfolding dflags (Cast con_app (mkSymCo co))
_ -> env1
; traceSmpl "addAltUnf" (vcat [ppr case_bndr <+> ppr scrut, ppr con_app])
; return env2 }
addBinderUnfolding :: SimplEnv -> Id -> Unfolding -> SimplEnv
addBinderUnfolding env bndr unf
| debugIsOn, Just tmpl <- maybeUnfoldingTemplate unf
= WARN( not (eqType (idType bndr) (exprType tmpl)),
ppr bndr $$ ppr (idType bndr) $$ ppr tmpl $$ ppr (exprType tmpl) )
modifyInScope env (bndr `setIdUnfolding` unf)
| otherwise
= modifyInScope env (bndr `setIdUnfolding` unf)
zapBndrOccInfo :: Bool -> Id -> Id
-- Consider case e of b { (a,b) -> ... }
-- Then if we bind b to (a,b) in "...", and b is not dead,
-- then we must zap the deadness info on a,b
zapBndrOccInfo keep_occ_info pat_id
| keep_occ_info = pat_id
| otherwise = zapIdOccInfo pat_id
{-
Note [Add unfolding for scrutinee]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
In general it's unlikely that a variable scrutinee will appear
in the case alternatives case x of { ...x unlikely to appear... }
because the binder-swap in OccAnal has got rid of all such occcurrences
See Note [Binder swap] in OccAnal.
BUT it is still VERY IMPORTANT to add a suitable unfolding for a
variable scrutinee, in simplAlt. Here's why
case x of y
(a,b) -> case b of c
I# v -> ...(f y)...
There is no occurrence of 'b' in the (...(f y)...). But y gets
the unfolding (a,b), and *that* mentions b. If f has a RULE
RULE f (p, I# q) = ...
we want that rule to match, so we must extend the in-scope env with a
suitable unfolding for 'y'. It's *essential* for rule matching; but
it's also good for case-elimintation -- suppose that 'f' was inlined
and did multi-level case analysis, then we'd solve it in one
simplifier sweep instead of two.
Exactly the same issue arises in SpecConstr;
see Note [Add scrutinee to ValueEnv too] in SpecConstr
HOWEVER, given
case x of y { Just a -> r1; Nothing -> r2 }
we do not want to add the unfolding x -> y to 'x', which might seem cool,
since 'y' itself has different unfoldings in r1 and r2. Reason: if we
did that, we'd have to zap y's deadness info and that is a very useful
piece of information.
So instead we add the unfolding x -> Just a, and x -> Nothing in the
respective RHSs.
************************************************************************
* *
\subsection{Known constructor}
* *
************************************************************************
We are a bit careful with occurrence info. Here's an example
(\x* -> case x of (a*, b) -> f a) (h v, e)
where the * means "occurs once". This effectively becomes
case (h v, e) of (a*, b) -> f a)
and then
let a* = h v; b = e in f a
and then
f (h v)
All this should happen in one sweep.
-}
knownCon :: SimplEnv
-> OutExpr -- The scrutinee
-> DataCon -> [OutType] -> [OutExpr] -- The scrutinee (in pieces)
-> InId -> [InBndr] -> InExpr -- The alternative
-> SimplCont
-> SimplM (SimplEnv, OutExpr)
knownCon env scrut dc dc_ty_args dc_args bndr bs rhs cont
= do { env' <- bind_args env bs dc_args
; env'' <- bind_case_bndr env'
; simplExprF env'' rhs cont }
where
zap_occ = zapBndrOccInfo (isDeadBinder bndr) -- bndr is an InId
-- Ugh!
bind_args env' [] _ = return env'
bind_args env' (b:bs') (Type ty : args)
= ASSERT( isTyVar b )
bind_args (extendTvSubst env' b ty) bs' args
bind_args env' (b:bs') (Coercion co : args)
= ASSERT( isCoVar b )
bind_args (extendCvSubst env' b co) bs' args
bind_args env' (b:bs') (arg : args)
= ASSERT( isId b )
do { let b' = zap_occ b
-- Note that the binder might be "dead", because it doesn't
-- occur in the RHS; and simplNonRecX may therefore discard
-- it via postInlineUnconditionally.
-- Nevertheless we must keep it if the case-binder is alive,
-- because it may be used in the con_app. See Note [knownCon occ info]
; env'' <- simplNonRecX env' b' arg -- arg satisfies let/app invariant
; bind_args env'' bs' args }
bind_args _ _ _ =
pprPanic "bind_args" $ ppr dc $$ ppr bs $$ ppr dc_args $$
text "scrut:" <+> ppr scrut
-- It's useful to bind bndr to scrut, rather than to a fresh
-- binding x = Con arg1 .. argn
-- because very often the scrut is a variable, so we avoid
-- creating, and then subsequently eliminating, a let-binding
-- BUT, if scrut is a not a variable, we must be careful
-- about duplicating the arg redexes; in that case, make
-- a new con-app from the args
bind_case_bndr env
| isDeadBinder bndr = return env
| exprIsTrivial scrut = return (extendIdSubst env bndr (DoneEx scrut))
| otherwise = do { dc_args <- mapM (simplVar env) bs
-- dc_ty_args are aready OutTypes,
-- but bs are InBndrs
; let con_app = Var (dataConWorkId dc)
`mkTyApps` dc_ty_args
`mkApps` dc_args
; simplNonRecX env bndr con_app }
-------------------
missingAlt :: SimplEnv -> Id -> [InAlt] -> SimplCont -> SimplM (SimplEnv, OutExpr)
-- This isn't strictly an error, although it is unusual.
-- It's possible that the simplifer might "see" that
-- an inner case has no accessible alternatives before
-- it "sees" that the entire branch of an outer case is
-- inaccessible. So we simply put an error case here instead.
missingAlt env case_bndr _ cont
= WARN( True, text "missingAlt" <+> ppr case_bndr )
return (env, mkImpossibleExpr (contResultType cont))
{-
************************************************************************
* *
\subsection{Duplicating continuations}
* *
************************************************************************
-}
prepareCaseCont :: SimplEnv
-> [InAlt] -> SimplCont
-> SimplM (SimplEnv,
SimplCont, -- Dupable part
SimplCont) -- Non-dupable part
-- We are considering
-- K[case _ of { p1 -> r1; ...; pn -> rn }]
-- where K is some enclosing continuation for the case
-- Goal: split K into two pieces Kdup,Knodup so that
-- a) Kdup can be duplicated
-- b) Knodup[Kdup[e]] = K[e]
-- The idea is that we'll transform thus:
-- Knodup[ (case _ of { p1 -> Kdup[r1]; ...; pn -> Kdup[rn] }
--
-- We may also return some extra bindings in SimplEnv (that scope over
-- the entire continuation)
--
-- When case-of-case is off, just make the entire continuation non-dupable
prepareCaseCont env alts cont
| not (sm_case_case (getMode env)) = return (env, mkBoringStop (contHoleType cont), cont)
| not (many_alts alts) = return (env, cont, mkBoringStop (contResultType cont))
| otherwise = mkDupableCont env cont
where
many_alts :: [InAlt] -> Bool -- True iff strictly > 1 non-bottom alternative
many_alts [] = False -- See Note [Bottom alternatives]
many_alts [_] = False
many_alts (alt:alts)
| is_bot_alt alt = many_alts alts
| otherwise = not (all is_bot_alt alts)
is_bot_alt (_,_,rhs) = exprIsBottom rhs
{-
Note [Bottom alternatives]
~~~~~~~~~~~~~~~~~~~~~~~~~~
When we have
case (case x of { A -> error .. ; B -> e; C -> error ..)
of alts
then we can just duplicate those alts because the A and C cases
will disappear immediately. This is more direct than creating
join points and inlining them away; and in some cases we would
not even create the join points (see Note [Single-alternative case])
and we would keep the case-of-case which is silly. See Trac #4930.
-}
mkDupableCont :: SimplEnv -> SimplCont
-> SimplM (SimplEnv, SimplCont, SimplCont)
mkDupableCont env cont
| contIsDupable cont
= return (env, cont, mkBoringStop (contResultType cont))
mkDupableCont _ (Stop {}) = panic "mkDupableCont" -- Handled by previous eqn
mkDupableCont env (CastIt ty cont)
= do { (env', dup, nodup) <- mkDupableCont env cont
; return (env', CastIt ty dup, nodup) }
-- Duplicating ticks for now, not sure if this is good or not
mkDupableCont env cont@(TickIt{})
= return (env, mkBoringStop (contHoleType cont), cont)
mkDupableCont env cont@(StrictBind {})
= return (env, mkBoringStop (contHoleType cont), cont)
-- See Note [Duplicating StrictBind]
mkDupableCont env (StrictArg info cci cont)
-- See Note [Duplicating StrictArg]
= do { (env', dup, nodup) <- mkDupableCont env cont
; (env'', args') <- mapAccumLM makeTrivialArg env' (ai_args info)
; return (env'', StrictArg (info { ai_args = args' }) cci dup, nodup) }
mkDupableCont env cont@(ApplyToTy { sc_cont = tail })
= do { (env', dup_cont, nodup_cont) <- mkDupableCont env tail
; return (env', cont { sc_cont = dup_cont }, nodup_cont ) }
mkDupableCont env (ApplyToVal { sc_arg = arg, sc_dup = dup, sc_env = se, sc_cont = cont })
= -- e.g. [...hole...] (...arg...)
-- ==>
-- let a = ...arg...
-- in [...hole...] a
do { (env', dup_cont, nodup_cont) <- mkDupableCont env cont
; (_, se', arg') <- simplArg env' dup se arg
; (env'', arg'') <- makeTrivial NotTopLevel env' (fsLit "karg") arg'
; let app_cont = ApplyToVal { sc_arg = arg'', sc_env = se'
, sc_dup = OkToDup, sc_cont = dup_cont }
; return (env'', app_cont, nodup_cont) }
mkDupableCont env cont@(Select { sc_bndr = case_bndr, sc_alts = [(_, bs, _rhs)] })
-- See Note [Single-alternative case]
-- | not (exprIsDupable rhs && contIsDupable case_cont)
-- | not (isDeadBinder case_bndr)
| all isDeadBinder bs -- InIds
&& not (isUnliftedType (idType case_bndr))
-- Note [Single-alternative-unlifted]
= return (env, mkBoringStop (contHoleType cont), cont)
mkDupableCont env (Select { sc_bndr = case_bndr, sc_alts = alts
, sc_env = se, sc_cont = cont })
= -- e.g. (case [...hole...] of { pi -> ei })
-- ===>
-- let ji = \xij -> ei
-- in case [...hole...] of { pi -> ji xij }
do { tick (CaseOfCase case_bndr)
; (env', dup_cont, nodup_cont) <- prepareCaseCont env alts cont
-- NB: We call prepareCaseCont here. If there is only one
-- alternative, then dup_cont may be big, but that's ok
-- because we push it into the single alternative, and then
-- use mkDupableAlt to turn that simplified alternative into
-- a join point if it's too big to duplicate.
-- And this is important: see Note [Fusing case continuations]
; let alt_env = se `setInScope` env'
; (alt_env', case_bndr') <- simplBinder alt_env case_bndr
; alts' <- mapM (simplAlt alt_env' Nothing [] case_bndr' dup_cont) alts
-- Safe to say that there are no handled-cons for the DEFAULT case
-- NB: simplBinder does not zap deadness occ-info, so
-- a dead case_bndr' will still advertise its deadness
-- This is really important because in
-- case e of b { (# p,q #) -> ... }
-- b is always dead, and indeed we are not allowed to bind b to (# p,q #),
-- which might happen if e was an explicit unboxed pair and b wasn't marked dead.
-- In the new alts we build, we have the new case binder, so it must retain
-- its deadness.
-- NB: we don't use alt_env further; it has the substEnv for
-- the alternatives, and we don't want that
; (env'', alts'') <- mkDupableAlts env' case_bndr' alts'
; return (env'', -- Note [Duplicated env]
Select { sc_dup = OkToDup
, sc_bndr = case_bndr', sc_alts = alts''
, sc_env = zapSubstEnv env''
, sc_cont = mkBoringStop (contHoleType nodup_cont) },
nodup_cont) }
mkDupableAlts :: SimplEnv -> OutId -> [InAlt]
-> SimplM (SimplEnv, [InAlt])
-- Absorbs the continuation into the new alternatives
mkDupableAlts env case_bndr' the_alts
= go env the_alts
where
go env0 [] = return (env0, [])
go env0 (alt:alts)
= do { (env1, alt') <- mkDupableAlt env0 case_bndr' alt
; (env2, alts') <- go env1 alts
; return (env2, alt' : alts' ) }
mkDupableAlt :: SimplEnv -> OutId -> (AltCon, [CoreBndr], CoreExpr)
-> SimplM (SimplEnv, (AltCon, [CoreBndr], CoreExpr))
mkDupableAlt env case_bndr (con, bndrs', rhs') = do
dflags <- getDynFlags
if exprIsDupable dflags rhs' -- Note [Small alternative rhs]
then return (env, (con, bndrs', rhs'))
else
do { let rhs_ty' = exprType rhs'
scrut_ty = idType case_bndr
case_bndr_w_unf
= case con of
DEFAULT -> case_bndr
DataAlt dc -> setIdUnfolding case_bndr unf
where
-- See Note [Case binders and join points]
unf = mkInlineUnfolding Nothing rhs
rhs = mkConApp2 dc (tyConAppArgs scrut_ty) bndrs'
LitAlt {} -> WARN( True, text "mkDupableAlt"
<+> ppr case_bndr <+> ppr con )
case_bndr
-- The case binder is alive but trivial, so why has
-- it not been substituted away?
used_bndrs' | isDeadBinder case_bndr = filter abstract_over bndrs'
| otherwise = bndrs' ++ [case_bndr_w_unf]
abstract_over bndr
| isTyVar bndr = True -- Abstract over all type variables just in case
| otherwise = not (isDeadBinder bndr)
-- The deadness info on the new Ids is preserved by simplBinders
; (final_bndrs', final_args) -- Note [Join point abstraction]
<- if (any isId used_bndrs')
then return (used_bndrs', varsToCoreExprs used_bndrs')
else do { rw_id <- newId (fsLit "w") voidPrimTy
; return ([setOneShotLambda rw_id], [Var voidPrimId]) }
; join_bndr <- newId (fsLit "$j") (mkPiTypes final_bndrs' rhs_ty')
-- Note [Funky mkPiTypes]
; let -- We make the lambdas into one-shot-lambdas. The
-- join point is sure to be applied at most once, and doing so
-- prevents the body of the join point being floated out by
-- the full laziness pass
really_final_bndrs = map one_shot final_bndrs'
one_shot v | isId v = setOneShotLambda v
| otherwise = v
join_rhs = mkLams really_final_bndrs rhs'
join_arity = exprArity join_rhs
join_call = mkApps (Var join_bndr) final_args
; env' <- addPolyBind NotTopLevel env (NonRec (join_bndr `setIdArity` join_arity) join_rhs)
; return (env', (con, bndrs', join_call)) }
-- See Note [Duplicated env]
{-
Note [Fusing case continuations]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
It's important to fuse two successive case continuations when the
first has one alternative. That's why we call prepareCaseCont here.
Consider this, which arises from thunk splitting (see Note [Thunk
splitting] in WorkWrap):
let
x* = case (case v of {pn -> rn}) of
I# a -> I# a
in body
The simplifier will find
(Var v) with continuation
Select (pn -> rn) (
Select [I# a -> I# a] (
StrictBind body Stop
So we'll call mkDupableCont on
Select [I# a -> I# a] (StrictBind body Stop)
There is just one alternative in the first Select, so we want to
simplify the rhs (I# a) with continuation (StricgtBind body Stop)
Supposing that body is big, we end up with
let $j a = <let x = I# a in body>
in case v of { pn -> case rn of
I# a -> $j a }
This is just what we want because the rn produces a box that
the case rn cancels with.
See Trac #4957 a fuller example.
Note [Case binders and join points]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Consider this
case (case .. ) of c {
I# c# -> ....c....
If we make a join point with c but not c# we get
$j = \c -> ....c....
But if later inlining scrutines the c, thus
$j = \c -> ... case c of { I# y -> ... } ...
we won't see that 'c' has already been scrutinised. This actually
happens in the 'tabulate' function in wave4main, and makes a significant
difference to allocation.
An alternative plan is this:
$j = \c# -> let c = I# c# in ...c....
but that is bad if 'c' is *not* later scrutinised.
So instead we do both: we pass 'c' and 'c#' , and record in c's inlining
(a stable unfolding) that it's really I# c#, thus
$j = \c# -> \c[=I# c#] -> ...c....
Absence analysis may later discard 'c'.
NB: take great care when doing strictness analysis;
see Note [Lamba-bound unfoldings] in DmdAnal.
Also note that we can still end up passing stuff that isn't used. Before
strictness analysis we have
let $j x y c{=(x,y)} = (h c, ...)
in ...
After strictness analysis we see that h is strict, we end up with
let $j x y c{=(x,y)} = ($wh x y, ...)
and c is unused.
Note [Duplicated env]
~~~~~~~~~~~~~~~~~~~~~
Some of the alternatives are simplified, but have not been turned into a join point
So they *must* have an zapped subst-env. So we can't use completeNonRecX to
bind the join point, because it might to do PostInlineUnconditionally, and
we'd lose that when zapping the subst-env. We could have a per-alt subst-env,
but zapping it (as we do in mkDupableCont, the Select case) is safe, and
at worst delays the join-point inlining.
Note [Small alternative rhs]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~
It is worth checking for a small RHS because otherwise we
get extra let bindings that may cause an extra iteration of the simplifier to
inline back in place. Quite often the rhs is just a variable or constructor.
The Ord instance of Maybe in PrelMaybe.hs, for example, took several extra
iterations because the version with the let bindings looked big, and so wasn't
inlined, but after the join points had been inlined it looked smaller, and so
was inlined.
NB: we have to check the size of rhs', not rhs.
Duplicating a small InAlt might invalidate occurrence information
However, if it *is* dupable, we return the *un* simplified alternative,
because otherwise we'd need to pair it up with an empty subst-env....
but we only have one env shared between all the alts.
(Remember we must zap the subst-env before re-simplifying something).
Rather than do this we simply agree to re-simplify the original (small) thing later.
Note [Funky mkPiTypes]
~~~~~~~~~~~~~~~~~~~~~~
Notice the funky mkPiTypes. If the contructor has existentials
it's possible that the join point will be abstracted over
type variables as well as term variables.
Example: Suppose we have
data T = forall t. C [t]
Then faced with
case (case e of ...) of
C t xs::[t] -> rhs
We get the join point
let j :: forall t. [t] -> ...
j = /\t \xs::[t] -> rhs
in
case (case e of ...) of
C t xs::[t] -> j t xs
Note [Join point abstraction]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Join points always have at least one value argument,
for several reasons
* If we try to lift a primitive-typed something out
for let-binding-purposes, we will *caseify* it (!),
with potentially-disastrous strictness results. So
instead we turn it into a function: \v -> e
where v::Void#. The value passed to this function is void,
which generates (almost) no code.
* CPR. We used to say "&& isUnliftedType rhs_ty'" here, but now
we make the join point into a function whenever used_bndrs'
is empty. This makes the join-point more CPR friendly.
Consider: let j = if .. then I# 3 else I# 4
in case .. of { A -> j; B -> j; C -> ... }
Now CPR doesn't w/w j because it's a thunk, so
that means that the enclosing function can't w/w either,
which is a lose. Here's the example that happened in practice:
kgmod :: Int -> Int -> Int
kgmod x y = if x > 0 && y < 0 || x < 0 && y > 0
then 78
else 5
* Let-no-escape. We want a join point to turn into a let-no-escape
so that it is implemented as a jump, and one of the conditions
for LNE is that it's not updatable. In CoreToStg, see
Note [What is a non-escaping let]
* Floating. Since a join point will be entered once, no sharing is
gained by floating out, but something might be lost by doing
so because it might be allocated.
I have seen a case alternative like this:
True -> \v -> ...
It's a bit silly to add the realWorld dummy arg in this case, making
$j = \s v -> ...
True -> $j s
(the \v alone is enough to make CPR happy) but I think it's rare
There's a slight infelicity here: we pass the overall
case_bndr to all the join points if it's used in *any* RHS,
because we don't know its usage in each RHS separately
Note [Duplicating StrictArg]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~
The original plan had (where E is a big argument)
e.g. f E [..hole..]
==> let $j = \a -> f E a
in $j [..hole..]
But this is terrible! Here's an example:
&& E (case x of { T -> F; F -> T })
Now, && is strict so we end up simplifying the case with
an ArgOf continuation. If we let-bind it, we get
let $j = \v -> && E v
in simplExpr (case x of { T -> F; F -> T })
(ArgOf (\r -> $j r)
And after simplifying more we get
let $j = \v -> && E v
in case x of { T -> $j F; F -> $j T }
Which is a Very Bad Thing
What we do now is this
f E [..hole..]
==> let a = E
in f a [..hole..]
Now if the thing in the hole is a case expression (which is when
we'll call mkDupableCont), we'll push the function call into the
branches, which is what we want. Now RULES for f may fire, and
call-pattern specialisation. Here's an example from Trac #3116
go (n+1) (case l of
1 -> bs'
_ -> Chunk p fpc (o+1) (l-1) bs')
If we can push the call for 'go' inside the case, we get
call-pattern specialisation for 'go', which is *crucial* for
this program.
Here is the (&&) example:
&& E (case x of { T -> F; F -> T })
==> let a = E in
case x of { T -> && a F; F -> && a T }
Much better!
Notice that
* Arguments to f *after* the strict one are handled by
the ApplyToVal case of mkDupableCont. Eg
f [..hole..] E
* We can only do the let-binding of E because the function
part of a StrictArg continuation is an explicit syntax
tree. In earlier versions we represented it as a function
(CoreExpr -> CoreEpxr) which we couldn't take apart.
Do *not* duplicate StrictBind and StritArg continuations. We gain
nothing by propagating them into the expressions, and we do lose a
lot.
The desire not to duplicate is the entire reason that
mkDupableCont returns a pair of continuations.
Note [Duplicating StrictBind]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Unlike StrictArg, there doesn't seem anything to gain from
duplicating a StrictBind continuation, so we don't.
Note [Single-alternative cases]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
This case is just like the ArgOf case. Here's an example:
data T a = MkT !a
...(MkT (abs x))...
Then we get
case (case x of I# x' ->
case x' <# 0# of
True -> I# (negate# x')
False -> I# x') of y {
DEFAULT -> MkT y
Because the (case x) has only one alternative, we'll transform to
case x of I# x' ->
case (case x' <# 0# of
True -> I# (negate# x')
False -> I# x') of y {
DEFAULT -> MkT y
But now we do *NOT* want to make a join point etc, giving
case x of I# x' ->
let $j = \y -> MkT y
in case x' <# 0# of
True -> $j (I# (negate# x'))
False -> $j (I# x')
In this case the $j will inline again, but suppose there was a big
strict computation enclosing the orginal call to MkT. Then, it won't
"see" the MkT any more, because it's big and won't get duplicated.
And, what is worse, nothing was gained by the case-of-case transform.
So, in circumstances like these, we don't want to build join points
and push the outer case into the branches of the inner one. Instead,
don't duplicate the continuation.
When should we use this strategy? We should not use it on *every*
single-alternative case:
e.g. case (case ....) of (a,b) -> (# a,b #)
Here we must push the outer case into the inner one!
Other choices:
* Match [(DEFAULT,_,_)], but in the common case of Int,
the alternative-filling-in code turned the outer case into
case (...) of y { I# _ -> MkT y }
* Match on single alternative plus (not (isDeadBinder case_bndr))
Rationale: pushing the case inwards won't eliminate the construction.
But there's a risk of
case (...) of y { (a,b) -> let z=(a,b) in ... }
Now y looks dead, but it'll come alive again. Still, this
seems like the best option at the moment.
* Match on single alternative plus (all (isDeadBinder bndrs))
Rationale: this is essentially seq.
* Match when the rhs is *not* duplicable, and hence would lead to a
join point. This catches the disaster-case above. We can test
the *un-simplified* rhs, which is fine. It might get bigger or
smaller after simplification; if it gets smaller, this case might
fire next time round. NB also that we must test contIsDupable
case_cont *too, because case_cont might be big!
HOWEVER: I found that this version doesn't work well, because
we can get let x = case (...) of { small } in ...case x...
When x is inlined into its full context, we find that it was a bad
idea to have pushed the outer case inside the (...) case.
There is a cost to not doing case-of-case; see Trac #10626.
Note [Single-alternative-unlifted]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Here's another single-alternative where we really want to do case-of-case:
data Mk1 = Mk1 Int# | Mk2 Int#
M1.f =
\r [x_s74 y_s6X]
case
case y_s6X of tpl_s7m {
M1.Mk1 ipv_s70 -> ipv_s70;
M1.Mk2 ipv_s72 -> ipv_s72;
}
of
wild_s7c
{ __DEFAULT ->
case
case x_s74 of tpl_s7n {
M1.Mk1 ipv_s77 -> ipv_s77;
M1.Mk2 ipv_s79 -> ipv_s79;
}
of
wild1_s7b
{ __DEFAULT -> ==# [wild1_s7b wild_s7c];
};
};
So the outer case is doing *nothing at all*, other than serving as a
join-point. In this case we really want to do case-of-case and decide
whether to use a real join point or just duplicate the continuation:
let $j s7c = case x of
Mk1 ipv77 -> (==) s7c ipv77
Mk1 ipv79 -> (==) s7c ipv79
in
case y of
Mk1 ipv70 -> $j ipv70
Mk2 ipv72 -> $j ipv72
Hence: check whether the case binder's type is unlifted, because then
the outer case is *not* a seq.
************************************************************************
* *
Unfoldings
* *
************************************************************************
-}
simplLetUnfolding :: SimplEnv-> TopLevelFlag
-> InId
-> OutExpr
-> Unfolding -> SimplM Unfolding
simplLetUnfolding env top_lvl id new_rhs unf
| isStableUnfolding unf
= simplUnfolding env top_lvl id unf
| otherwise
= bottoming `seq` -- See Note [Force bottoming field]
do { dflags <- getDynFlags
; return (mkUnfolding dflags InlineRhs (isTopLevel top_lvl) bottoming new_rhs) }
-- We make an unfolding *even for loop-breakers*.
-- Reason: (a) It might be useful to know that they are WHNF
-- (b) In TidyPgm we currently assume that, if we want to
-- expose the unfolding then indeed we *have* an unfolding
-- to expose. (We could instead use the RHS, but currently
-- we don't.) The simple thing is always to have one.
where
bottoming = isBottomingId id
simplUnfolding :: SimplEnv-> TopLevelFlag -> InId -> Unfolding -> SimplM Unfolding
-- Note [Setting the new unfolding]
simplUnfolding env top_lvl id unf
= case unf of
NoUnfolding -> return unf
OtherCon {} -> return unf
DFunUnfolding { df_bndrs = bndrs, df_con = con, df_args = args }
-> do { (env', bndrs') <- simplBinders rule_env bndrs
; args' <- mapM (simplExpr env') args
; return (mkDFunUnfolding bndrs' con args') }
CoreUnfolding { uf_tmpl = expr, uf_src = src, uf_guidance = guide }
| isStableSource src
-> do { expr' <- simplExpr rule_env expr
; case guide of
UnfWhen { ug_arity = arity, ug_unsat_ok = sat_ok } -- Happens for INLINE things
-> let guide' = UnfWhen { ug_arity = arity, ug_unsat_ok = sat_ok
, ug_boring_ok = inlineBoringOk expr' }
-- Refresh the boring-ok flag, in case expr'
-- has got small. This happens, notably in the inlinings
-- for dfuns for single-method classes; see
-- Note [Single-method classes] in TcInstDcls.
-- A test case is Trac #4138
in return (mkCoreUnfolding src is_top_lvl expr' guide')
-- See Note [Top-level flag on inline rules] in CoreUnfold
_other -- Happens for INLINABLE things
-> bottoming `seq` -- See Note [Force bottoming field]
do { dflags <- getDynFlags
; return (mkUnfolding dflags src is_top_lvl bottoming expr') } }
-- If the guidance is UnfIfGoodArgs, this is an INLINABLE
-- unfolding, and we need to make sure the guidance is kept up
-- to date with respect to any changes in the unfolding.
| otherwise -> return noUnfolding -- Discard unstable unfoldings
where
bottoming = isBottomingId id
is_top_lvl = isTopLevel top_lvl
act = idInlineActivation id
rule_env = updMode (updModeForStableUnfoldings act) env
-- See Note [Simplifying inside stable unfoldings] in SimplUtils
{-
Note [Force bottoming field]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~
We need to force bottoming, or the new unfolding holds
on to the old unfolding (which is part of the id).
Note [Setting the new unfolding]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
* If there's an INLINE pragma, we simplify the RHS gently. Maybe we
should do nothing at all, but simplifying gently might get rid of
more crap.
* If not, we make an unfolding from the new RHS. But *only* for
non-loop-breakers. Making loop breakers not have an unfolding at all
means that we can avoid tests in exprIsConApp, for example. This is
important: if exprIsConApp says 'yes' for a recursive thing, then we
can get into an infinite loop
If there's an stable unfolding on a loop breaker (which happens for
INLINEABLE), we hang on to the inlining. It's pretty dodgy, but the
user did say 'INLINE'. May need to revisit this choice.
************************************************************************
* *
Rules
* *
************************************************************************
Note [Rules in a letrec]
~~~~~~~~~~~~~~~~~~~~~~~~
After creating fresh binders for the binders of a letrec, we
substitute the RULES and add them back onto the binders; this is done
*before* processing any of the RHSs. This is important. Manuel found
cases where he really, really wanted a RULE for a recursive function
to apply in that function's own right-hand side.
See Note [Loop breaking and RULES] in OccAnal.
-}
addBndrRules :: SimplEnv -> InBndr -> OutBndr -> SimplM (SimplEnv, OutBndr)
-- Rules are added back into the bin
addBndrRules env in_id out_id
| null old_rules
= return (env, out_id)
| otherwise
= do { new_rules <- simplRules env (Just (idName out_id)) old_rules
; let final_id = out_id `setIdSpecialisation` mkRuleInfo new_rules
; return (modifyInScope env final_id, final_id) }
where
old_rules = ruleInfoRules (idSpecialisation in_id)
simplRules :: SimplEnv -> Maybe Name -> [CoreRule] -> SimplM [CoreRule]
simplRules env mb_new_nm rules
= mapM simpl_rule rules
where
simpl_rule rule@(BuiltinRule {})
= return rule
simpl_rule rule@(Rule { ru_bndrs = bndrs, ru_args = args
, ru_fn = fn_name, ru_rhs = rhs })
= do { (env', bndrs') <- simplBinders env bndrs
; let rule_env = updMode updModeForRules env'
; args' <- mapM (simplExpr rule_env) args
; rhs' <- simplExpr rule_env rhs
; return (rule { ru_bndrs = bndrs'
, ru_fn = mb_new_nm `orElse` fn_name
, ru_args = args'
, ru_rhs = rhs' }) }
| tjakway/ghcjvm | compiler/simplCore/Simplify.hs | bsd-3-clause | 124,411 | 20 | 25 | 38,197 | 15,479 | 8,190 | 7,289 | -1 | -1 |
{-# LANGUAGE OverloadedStrings #-}
module Distribution.Parsec.ConfVar (parseConditionConfVar) where
import Prelude ()
import Distribution.Compat.Prelude
import Distribution.Compat.Parsec (integral)
import Distribution.Parsec.Class (Parsec (..))
import Distribution.Parsec.Types.Common
import Distribution.Parsec.Types.Field (SectionArg (..))
import Distribution.Parsec.Types.ParseResult
import Distribution.Simple.Utils (fromUTF8BS)
import Distribution.Types.Condition
import Distribution.Types.GenericPackageDescription
(ConfVar (..))
import Distribution.Version
(anyVersion, earlierVersion, intersectVersionRanges,
laterVersion, majorBoundVersion, mkVersion, noVersion,
orEarlierVersion, orLaterVersion, thisVersion,
unionVersionRanges, withinVersion)
import qualified Text.Parsec as P
import qualified Text.Parsec.Error as P
-- | Parse @'Condition' 'ConfVar'@ from section arguments provided by parsec
-- based outline parser.
parseConditionConfVar :: [SectionArg Position] -> ParseResult (Condition ConfVar)
parseConditionConfVar args = do
-- preprocess glued operators
args' <- preprocess args
-- The name of the input file is irrelevant, as we reformat the error message.
case P.runParser (parser <* P.eof) () "<condition>" args' of
Right x -> pure x
Left err -> do
-- Mangle the position to the actual one
let ppos = P.errorPos err
let epos = Position (P.sourceLine ppos) (P.sourceColumn ppos)
let msg = P.showErrorMessages
"or" "unknown parse error" "expecting" "unexpected" "end of input"
(P.errorMessages err)
parseFailure epos msg
pure $ Lit True
-- This is a hack, as we have "broken" .cabal files on Hackage
--
-- There are glued operators "&&!" (no whitespace) in some cabal files.
-- E.g. http://hackage.haskell.org/package/hblas-0.2.0.0/hblas.cabal
preprocess :: [SectionArg Position] -> ParseResult [SectionArg Position]
preprocess (SecArgOther pos "&&!" : rest) = do
parseWarning pos PWTGluedOperators "Glued operators: &&!"
(\rest' -> SecArgOther pos "&&" : SecArgOther pos "!" : rest') <$> preprocess rest
preprocess (x : rest) =
(x: ) <$> preprocess rest
preprocess [] = pure []
type Parser = P.Parsec [SectionArg Position] ()
parser :: Parser (Condition ConfVar)
parser = condOr
where
condOr = P.sepBy1 condAnd (oper "||") >>= return . foldl1 COr
condAnd = P.sepBy1 cond (oper "&&") >>= return . foldl1 CAnd
cond = P.choice
[ boolLiteral, parens condOr, notCond, osCond, archCond, flagCond, implCond ]
notCond = CNot <$ oper "!" <*> cond
boolLiteral = Lit <$> boolLiteral'
osCond = Var . OS <$ string "os" <*> parens fromParsec
flagCond = Var . Flag <$ string "flag" <*> parens fromParsec
archCond = Var . Arch <$ string "arch" <*> parens fromParsec
implCond = Var <$ string "impl" <*> parens implCond'
implCond' = Impl
<$> fromParsec
<*> P.option anyVersion versionRange
version = fromParsec
versionStar = mkVersion <$> fromParsec' versionStar' <* oper "*"
versionStar' = some (integral <* P.char '.')
versionRange = expr
where
expr = foldl1 unionVersionRanges <$> P.sepBy1 term (oper "||")
term = foldl1 intersectVersionRanges <$> P.sepBy1 factor (oper "&&")
factor = P.choice
$ parens expr
: parseAnyVersion
: parseNoVersion
: parseWildcardRange
: map parseRangeOp rangeOps
parseAnyVersion = anyVersion <$ string "-any"
parseNoVersion = noVersion <$ string "-none"
parseWildcardRange = P.try $ withinVersion <$ oper "==" <*> versionStar
parseRangeOp (s,f) = P.try (f <$ oper s <*> version)
rangeOps = [ ("<", earlierVersion),
("<=", orEarlierVersion),
(">", laterVersion),
(">=", orLaterVersion),
("^>=", majorBoundVersion),
("==", thisVersion) ]
-- Number token can have many dots in it: SecArgNum (Position 65 15) "7.6.1"
ident = tokenPrim $ \t -> case t of
SecArgName _ s -> Just $ fromUTF8BS s
SecArgNum _ s -> Just $ fromUTF8BS s
_ -> Nothing
boolLiteral' = tokenPrim $ \t -> case t of
SecArgName _ s
| s == "True" -> Just True
| s == "true" -> Just True
| s == "False" -> Just False
| s == "false" -> Just False
_ -> Nothing
string s = tokenPrim $ \t -> case t of
SecArgName _ s' | s == s' -> Just ()
_ -> Nothing
oper o = tokenPrim $ \t -> case t of
SecArgOther _ o' | o == o' -> Just ()
_ -> Nothing
parens = P.between (oper "(") (oper ")")
tokenPrim = P.tokenPrim prettySectionArg updatePosition
-- TODO: check where the errors are reported
updatePosition x _ _ = x
prettySectionArg = show
fromParsec :: Parsec a => Parser a
fromParsec = fromParsec' parsec
fromParsec' p = do
i <- ident
case P.runParser (p <* P.eof) [] "<ident>" i of
Right x -> pure x
-- TODO: better lifting or errors / warnings
Left err -> fail $ show err
| mydaum/cabal | Cabal/Distribution/Parsec/ConfVar.hs | bsd-3-clause | 5,766 | 0 | 18 | 1,885 | 1,453 | 753 | 700 | 106 | 7 |
module One where
import qualified Two
resource = "This is the sub-plugin of (" ++ Two.resource ++ ")"
| abuiles/turbinado-blog | tmp/dependencies/hs-plugins-1.3.1/testsuite/hier/hier3/One.hs | bsd-3-clause | 106 | 0 | 7 | 22 | 23 | 14 | 9 | 3 | 1 |
{-# LANGUAGE PolyKinds #-}
module T16456 where
data T p = MkT
foo :: T Int
foo = _
| sdiehl/ghc | testsuite/tests/typecheck/should_fail/T16456.hs | bsd-3-clause | 85 | 0 | 5 | 21 | 26 | 16 | 10 | 5 | 1 |
{-
(c) The University of Glasgow 2006
(c) The GRASP/AQUA Project, Glasgow University, 1992-1998
\section[NameEnv]{@NameEnv@: name environments}
-}
{-# LANGUAGE CPP #-}
module NameEnv (
-- * Var, Id and TyVar environments (maps)
NameEnv,
-- ** Manipulating these environments
mkNameEnv,
emptyNameEnv, isEmptyNameEnv,
unitNameEnv, nameEnvElts, nameEnvUniqueElts,
extendNameEnv_C, extendNameEnv_Acc, extendNameEnv,
extendNameEnvList, extendNameEnvList_C,
foldNameEnv, filterNameEnv, anyNameEnv,
plusNameEnv, plusNameEnv_C, alterNameEnv,
lookupNameEnv, lookupNameEnv_NF, delFromNameEnv, delListFromNameEnv,
elemNameEnv, mapNameEnv, disjointNameEnv,
-- ** Dependency analysis
depAnal
) where
#include "HsVersions.h"
import Digraph
import Name
import Unique
import UniqFM
import Maybes
{-
************************************************************************
* *
\subsection{Name environment}
* *
************************************************************************
-}
depAnal :: (node -> [Name]) -- Defs
-> (node -> [Name]) -- Uses
-> [node]
-> [SCC node]
-- Peform dependency analysis on a group of definitions,
-- where each definition may define more than one Name
--
-- The get_defs and get_uses functions are called only once per node
depAnal get_defs get_uses nodes
= stronglyConnCompFromEdgedVertices (map mk_node keyed_nodes)
where
keyed_nodes = nodes `zip` [(1::Int)..]
mk_node (node, key) = (node, key, mapMaybe (lookupNameEnv key_map) (get_uses node))
key_map :: NameEnv Int -- Maps a Name to the key of the decl that defines it
key_map = mkNameEnv [(name,key) | (node, key) <- keyed_nodes, name <- get_defs node]
{-
************************************************************************
* *
\subsection{Name environment}
* *
************************************************************************
-}
type NameEnv a = UniqFM a -- Domain is Name
emptyNameEnv :: NameEnv a
isEmptyNameEnv :: NameEnv a -> Bool
mkNameEnv :: [(Name,a)] -> NameEnv a
nameEnvElts :: NameEnv a -> [a]
nameEnvUniqueElts :: NameEnv a -> [(Unique, a)]
alterNameEnv :: (Maybe a-> Maybe a) -> NameEnv a -> Name -> NameEnv a
extendNameEnv_C :: (a->a->a) -> NameEnv a -> Name -> a -> NameEnv a
extendNameEnv_Acc :: (a->b->b) -> (a->b) -> NameEnv b -> Name -> a -> NameEnv b
extendNameEnv :: NameEnv a -> Name -> a -> NameEnv a
plusNameEnv :: NameEnv a -> NameEnv a -> NameEnv a
plusNameEnv_C :: (a->a->a) -> NameEnv a -> NameEnv a -> NameEnv a
extendNameEnvList :: NameEnv a -> [(Name,a)] -> NameEnv a
extendNameEnvList_C :: (a->a->a) -> NameEnv a -> [(Name,a)] -> NameEnv a
delFromNameEnv :: NameEnv a -> Name -> NameEnv a
delListFromNameEnv :: NameEnv a -> [Name] -> NameEnv a
elemNameEnv :: Name -> NameEnv a -> Bool
unitNameEnv :: Name -> a -> NameEnv a
lookupNameEnv :: NameEnv a -> Name -> Maybe a
lookupNameEnv_NF :: NameEnv a -> Name -> a
foldNameEnv :: (a -> b -> b) -> b -> NameEnv a -> b
filterNameEnv :: (elt -> Bool) -> NameEnv elt -> NameEnv elt
anyNameEnv :: (elt -> Bool) -> NameEnv elt -> Bool
mapNameEnv :: (elt1 -> elt2) -> NameEnv elt1 -> NameEnv elt2
disjointNameEnv :: NameEnv a -> NameEnv a -> Bool
nameEnvElts x = eltsUFM x
emptyNameEnv = emptyUFM
isEmptyNameEnv = isNullUFM
unitNameEnv x y = unitUFM x y
extendNameEnv x y z = addToUFM x y z
extendNameEnvList x l = addListToUFM x l
lookupNameEnv x y = lookupUFM x y
alterNameEnv = alterUFM
mkNameEnv l = listToUFM l
elemNameEnv x y = elemUFM x y
foldNameEnv a b c = foldUFM a b c
plusNameEnv x y = plusUFM x y
plusNameEnv_C f x y = plusUFM_C f x y
extendNameEnv_C f x y z = addToUFM_C f x y z
mapNameEnv f x = mapUFM f x
nameEnvUniqueElts x = ufmToList x
extendNameEnv_Acc x y z a b = addToUFM_Acc x y z a b
extendNameEnvList_C x y z = addListToUFM_C x y z
delFromNameEnv x y = delFromUFM x y
delListFromNameEnv x y = delListFromUFM x y
filterNameEnv x y = filterUFM x y
anyNameEnv f x = foldUFM ((||) . f) False x
disjointNameEnv x y = isNullUFM (intersectUFM x y)
lookupNameEnv_NF env n = expectJust "lookupNameEnv_NF" (lookupNameEnv env n)
| tjakway/ghcjvm | compiler/basicTypes/NameEnv.hs | bsd-3-clause | 4,708 | 0 | 11 | 1,328 | 1,271 | 666 | 605 | 77 | 1 |
<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE helpset PUBLIC "-//Sun Microsystems Inc.//DTD JavaHelp HelpSet Version 2.0//EN" "http://java.sun.com/products/javahelp/helpset_2_0.dtd">
<helpset version="2.0" xml:lang="zh-CN">
<title>Groovy Support</title>
<maps>
<homeID>top</homeID>
<mapref location="map.jhm"/>
</maps>
<view>
<name>TOC</name>
<label>Contents</label>
<type>org.zaproxy.zap.extension.help.ZapTocView</type>
<data>toc.xml</data>
</view>
<view>
<name>Index</name>
<label>Index</label>
<type>javax.help.IndexView</type>
<data>index.xml</data>
</view>
<view>
<name>Search</name>
<label>Search</label>
<type>javax.help.SearchView</type>
<data engine="com.sun.java.help.search.DefaultSearchEngine">
JavaHelpSearch
</data>
</view>
<view>
<name>Favorites</name>
<label>Favorites</label>
<type>javax.help.FavoritesView</type>
</view>
</helpset> | thc202/zap-extensions | addOns/groovy/src/main/javahelp/org/zaproxy/zap/extension/groovy/resources/help_zh_CN/helpset_zh_CN.hs | apache-2.0 | 959 | 82 | 52 | 156 | 390 | 206 | 184 | -1 | -1 |
module MultiParamIn3 where
fromMaybe :: (Maybe a) -> a
fromMaybe (Just x) = x
fromMaybe Nothing = error "fromMaybe: Nothing"
f :: (Maybe Int) -> [Int] -> (Either Int b) -> Int
f Nothing y@[] (Left a) = (hd y) + a
f Nothing y@(b_1 : b_2) (Left a) = (hd y) + a
f (Just x) y@[] (Right b) = (hd y) + (fromMaybe x)
f (Just x) y@(b_1 : b_2) (Right b)
= (hd y) + (fromMaybe x)
f Nothing y (Left a) = (hd y) + a
f (Just x) y (Right b) = (hd y) + (fromMaybe x)
hd x = head x
tl x = tail x
| kmate/HaRe | old/testing/introPattern/MultiParamIn3AST.hs | bsd-3-clause | 495 | 0 | 9 | 127 | 334 | 171 | 163 | 14 | 1 |
{-# LANGUAGE Trustworthy #-}
{-# LANGUAGE NoImplicitPrelude, ExistentialQuantification #-}
-----------------------------------------------------------------------------
-- |
-- Module : Control.Exception
-- Copyright : (c) The University of Glasgow 2001
-- License : BSD-style (see the file libraries/base/LICENSE)
--
-- Maintainer : [email protected]
-- Stability : experimental
-- Portability : non-portable (extended exceptions)
--
-- This module provides support for raising and catching both built-in
-- and user-defined exceptions.
--
-- In addition to exceptions thrown by 'IO' operations, exceptions may
-- be thrown by pure code (imprecise exceptions) or by external events
-- (asynchronous exceptions), but may only be caught in the 'IO' monad.
-- For more details, see:
--
-- * /A semantics for imprecise exceptions/, by Simon Peyton Jones,
-- Alastair Reid, Tony Hoare, Simon Marlow, Fergus Henderson,
-- in /PLDI'99/.
--
-- * /Asynchronous exceptions in Haskell/, by Simon Marlow, Simon Peyton
-- Jones, Andy Moran and John Reppy, in /PLDI'01/.
--
-- * /An Extensible Dynamically-Typed Hierarchy of Exceptions/,
-- by Simon Marlow, in /Haskell '06/.
--
-----------------------------------------------------------------------------
module Control.Exception (
-- * The Exception type
SomeException(..),
Exception(..), -- class
IOException, -- instance Eq, Ord, Show, Typeable, Exception
ArithException(..), -- instance Eq, Ord, Show, Typeable, Exception
ArrayException(..), -- instance Eq, Ord, Show, Typeable, Exception
AssertionFailed(..),
SomeAsyncException(..),
AsyncException(..), -- instance Eq, Ord, Show, Typeable, Exception
asyncExceptionToException, asyncExceptionFromException,
NonTermination(..),
NestedAtomically(..),
BlockedIndefinitelyOnMVar(..),
BlockedIndefinitelyOnSTM(..),
AllocationLimitExceeded(..),
CompactionFailed(..),
Deadlock(..),
NoMethodError(..),
PatternMatchFail(..),
RecConError(..),
RecSelError(..),
RecUpdError(..),
ErrorCall(..),
TypeError(..),
-- * Throwing exceptions
throw,
throwIO,
ioError,
throwTo,
-- * Catching Exceptions
-- $catching
-- ** Catching all exceptions
-- $catchall
-- ** The @catch@ functions
catch,
catches, Handler(..),
catchJust,
-- ** The @handle@ functions
handle,
handleJust,
-- ** The @try@ functions
try,
tryJust,
-- ** The @evaluate@ function
evaluate,
-- ** The @mapException@ function
mapException,
-- * Asynchronous Exceptions
-- $async
-- ** Asynchronous exception control
-- |The following functions allow a thread to control delivery of
-- asynchronous exceptions during a critical region.
mask,
mask_,
uninterruptibleMask,
uninterruptibleMask_,
MaskingState(..),
getMaskingState,
interruptible,
allowInterrupt,
-- *** Applying @mask@ to an exception handler
-- $block_handler
-- *** Interruptible operations
-- $interruptible
-- * Assertions
assert,
-- * Utilities
bracket,
bracket_,
bracketOnError,
finally,
onException,
) where
import Control.Exception.Base
import GHC.Base
import GHC.IO (interruptible)
-- | You need this when using 'catches'.
data Handler a = forall e . Exception e => Handler (e -> IO a)
-- | @since 4.6.0.0
instance Functor Handler where
fmap f (Handler h) = Handler (fmap f . h)
{- |
Sometimes you want to catch two different sorts of exception. You could
do something like
> f = expr `catch` \ (ex :: ArithException) -> handleArith ex
> `catch` \ (ex :: IOException) -> handleIO ex
However, there are a couple of problems with this approach. The first is
that having two exception handlers is inefficient. However, the more
serious issue is that the second exception handler will catch exceptions
in the first, e.g. in the example above, if @handleArith@ throws an
@IOException@ then the second exception handler will catch it.
Instead, we provide a function 'catches', which would be used thus:
> f = expr `catches` [Handler (\ (ex :: ArithException) -> handleArith ex),
> Handler (\ (ex :: IOException) -> handleIO ex)]
-}
catches :: IO a -> [Handler a] -> IO a
catches io handlers = io `catch` catchesHandler handlers
catchesHandler :: [Handler a] -> SomeException -> IO a
catchesHandler handlers e = foldr tryHandler (throw e) handlers
where tryHandler (Handler handler) res
= case fromException e of
Just e' -> handler e'
Nothing -> res
-- -----------------------------------------------------------------------------
-- Catching exceptions
{- $catching
There are several functions for catching and examining
exceptions; all of them may only be used from within the
'IO' monad.
Here's a rule of thumb for deciding which catch-style function to
use:
* If you want to do some cleanup in the event that an exception
is raised, use 'finally', 'bracket' or 'onException'.
* To recover after an exception and do something else, the best
choice is to use one of the 'try' family.
* ... unless you are recovering from an asynchronous exception, in which
case use 'catch' or 'catchJust'.
The difference between using 'try' and 'catch' for recovery is that in
'catch' the handler is inside an implicit 'mask' (see \"Asynchronous
Exceptions\") which is important when catching asynchronous
exceptions, but when catching other kinds of exception it is
unnecessary. Furthermore it is possible to accidentally stay inside
the implicit 'mask' by tail-calling rather than returning from the
handler, which is why we recommend using 'try' rather than 'catch' for
ordinary exception recovery.
A typical use of 'tryJust' for recovery looks like this:
> do r <- tryJust (guard . isDoesNotExistError) $ getEnv "HOME"
> case r of
> Left e -> ...
> Right home -> ...
-}
-- -----------------------------------------------------------------------------
-- Asynchronous exceptions
-- | When invoked inside 'mask', this function allows a masked
-- asynchronous exception to be raised, if one exists. It is
-- equivalent to performing an interruptible operation (see
-- #interruptible), but does not involve any actual blocking.
--
-- When called outside 'mask', or inside 'uninterruptibleMask', this
-- function has no effect.
--
-- @since 4.4.0.0
allowInterrupt :: IO ()
allowInterrupt = interruptible $ return ()
{- $async
#AsynchronousExceptions# Asynchronous exceptions are so-called because they arise due to
external influences, and can be raised at any point during execution.
'StackOverflow' and 'HeapOverflow' are two examples of
system-generated asynchronous exceptions.
The primary source of asynchronous exceptions, however, is
'throwTo':
> throwTo :: ThreadId -> Exception -> IO ()
'throwTo' (also 'Control.Concurrent.killThread') allows one
running thread to raise an arbitrary exception in another thread. The
exception is therefore asynchronous with respect to the target thread,
which could be doing anything at the time it receives the exception.
Great care should be taken with asynchronous exceptions; it is all too
easy to introduce race conditions by the over zealous use of
'throwTo'.
-}
{- $block_handler
There\'s an implied 'mask' around every exception handler in a call
to one of the 'catch' family of functions. This is because that is
what you want most of the time - it eliminates a common race condition
in starting an exception handler, because there may be no exception
handler on the stack to handle another exception if one arrives
immediately. If asynchronous exceptions are masked on entering the
handler, though, we have time to install a new exception handler
before being interrupted. If this weren\'t the default, one would have
to write something like
> mask $ \restore ->
> catch (restore (...))
> (\e -> handler)
If you need to unmask asynchronous exceptions again in the exception
handler, 'restore' can be used there too.
Note that 'try' and friends /do not/ have a similar default, because
there is no exception handler in this case. Don't use 'try' for
recovering from an asynchronous exception.
-}
{- $interruptible
#interruptible#
Some operations are /interruptible/, which means that they can receive
asynchronous exceptions even in the scope of a 'mask'. Any function
which may itself block is defined as interruptible; this includes
'Control.Concurrent.MVar.takeMVar'
(but not 'Control.Concurrent.MVar.tryTakeMVar'),
and most operations which perform
some I\/O with the outside world. The reason for having
interruptible operations is so that we can write things like
> mask $ \restore -> do
> a <- takeMVar m
> catch (restore (...))
> (\e -> ...)
if the 'Control.Concurrent.MVar.takeMVar' was not interruptible,
then this particular
combination could lead to deadlock, because the thread itself would be
blocked in a state where it can\'t receive any asynchronous exceptions.
With 'Control.Concurrent.MVar.takeMVar' interruptible, however, we can be
safe in the knowledge that the thread can receive exceptions right up
until the point when the 'Control.Concurrent.MVar.takeMVar' succeeds.
Similar arguments apply for other interruptible operations like
'System.IO.openFile'.
It is useful to think of 'mask' not as a way to completely prevent
asynchronous exceptions, but as a way to switch from asynchronous mode
to polling mode. The main difficulty with asynchronous
exceptions is that they normally can occur anywhere, but within a
'mask' an asynchronous exception is only raised by operations that are
interruptible (or call other interruptible operations). In many cases
these operations may themselves raise exceptions, such as I\/O errors,
so the caller will usually be prepared to handle exceptions arising from the
operation anyway. To perform an explicit poll for asynchronous exceptions
inside 'mask', use 'allowInterrupt'.
Sometimes it is too onerous to handle exceptions in the middle of a
critical piece of stateful code. There are three ways to handle this
kind of situation:
* Use STM. Since a transaction is always either completely executed
or not at all, transactions are a good way to maintain invariants
over state in the presence of asynchronous (and indeed synchronous)
exceptions.
* Use 'mask', and avoid interruptible operations. In order to do
this, we have to know which operations are interruptible. It is
impossible to know for any given library function whether it might
invoke an interruptible operation internally; so instead we give a
list of guaranteed-not-to-be-interruptible operations below.
* Use 'uninterruptibleMask'. This is generally not recommended,
unless you can guarantee that any interruptible operations invoked
during the scope of 'uninterruptibleMask' can only ever block for
a short time. Otherwise, 'uninterruptibleMask' is a good way to
make your program deadlock and be unresponsive to user interrupts.
The following operations are guaranteed not to be interruptible:
* operations on 'IORef' from "Data.IORef"
* STM transactions that do not use 'retry'
* everything from the @Foreign@ modules
* everything from @Control.Exception@ except for 'throwTo'
* @tryTakeMVar@, @tryPutMVar@, @isEmptyMVar@
* @takeMVar@ if the @MVar@ is definitely full, and conversely @putMVar@ if the @MVar@ is definitely empty
* @newEmptyMVar@, @newMVar@
* @forkIO@, @forkIOUnmasked@, @myThreadId@
-}
{- $catchall
It is possible to catch all exceptions, by using the type 'SomeException':
> catch f (\e -> ... (e :: SomeException) ...)
HOWEVER, this is normally not what you want to do!
For example, suppose you want to read a file, but if it doesn't exist
then continue as if it contained \"\". You might be tempted to just
catch all exceptions and return \"\" in the handler. However, this has
all sorts of undesirable consequences. For example, if the user
presses control-C at just the right moment then the 'UserInterrupt'
exception will be caught, and the program will continue running under
the belief that the file contains \"\". Similarly, if another thread
tries to kill the thread reading the file then the 'ThreadKilled'
exception will be ignored.
Instead, you should only catch exactly the exceptions that you really
want. In this case, this would likely be more specific than even
\"any IO exception\"; a permissions error would likely also want to be
handled differently. Instead, you would probably want something like:
> e <- tryJust (guard . isDoesNotExistError) (readFile f)
> let str = either (const "") id e
There are occassions when you really do need to catch any sort of
exception. However, in most cases this is just so you can do some
cleaning up; you aren't actually interested in the exception itself.
For example, if you open a file then you want to close it again,
whether processing the file executes normally or throws an exception.
However, in these cases you can use functions like 'bracket', 'finally'
and 'onException', which never actually pass you the exception, but
just call the cleanup functions at the appropriate points.
But sometimes you really do need to catch any exception, and actually
see what the exception is. One example is at the very top-level of a
program, you may wish to catch any exception, print it to a logfile or
the screen, and then exit gracefully. For these cases, you can use
'catch' (or one of the other exception-catching functions) with the
'SomeException' type.
-}
| olsner/ghc | libraries/base/Control/Exception.hs | bsd-3-clause | 14,104 | 0 | 10 | 2,951 | 582 | 382 | 200 | 69 | 2 |
{-# LANGUAGE CPP #-}
module Examples.Commands where
import Data.List
import Options.Applicative
#if __GLASGOW_HASKELL__ <= 702
import Data.Monoid
(<>) :: Monoid a => a -> a -> a
(<>) = mappend
#endif
data Sample
= Hello [String]
| Goodbye
deriving Show
hello :: Parser Sample
hello = Hello <$> many (argument str (metavar "TARGET..."))
sample :: Parser Sample
sample = subparser
( command "hello"
(info hello
(progDesc "Print greeting"))
<> command "goodbye"
(info (pure Goodbye)
(progDesc "Say goodbye"))
)
run :: Sample -> IO ()
run (Hello targets) = putStrLn $ "Hello, " ++ intercalate ", " targets ++ "!"
run Goodbye = putStrLn "Goodbye."
opts :: ParserInfo Sample
opts = info (sample <**> helper) idm
main :: IO ()
main = execParser opts >>= run
| begriffs/optparse-applicative | tests/Examples/Commands.hs | bsd-3-clause | 833 | 0 | 12 | 207 | 275 | 145 | 130 | 28 | 1 |
module ShouldCompile where
-- I bet this test is a mistake! From the layout it
-- looks as if 'test' takes three args, the latter two
-- of higher rank. But the parens around these args are
-- missing, so it parses as
-- test :: [a]
-- -> forall a. Ord a
-- => [b]
-- -> forall c. Num c
-- => [c]
-- -> [a]
--
-- But maybe that what was intended; I'm not sure
-- Anyway it should typecheck!
test :: [a] -- ^ doc1
-> forall b. (Ord b) => [b] {-^ doc2 -}
-> forall c. (Num c) => [c] -- ^ doc3
-> [a]
test xs ys zs = xs
| spacekitteh/smcghc | testsuite/tests/haddock/should_compile_noflag_haddock/haddockC027.hs | bsd-3-clause | 599 | 0 | 12 | 200 | 89 | 58 | 31 | -1 | -1 |
{-# OPTIONS_GHC -fwarn-incomplete-patterns -fwarn-overlapping-patterns #-}
{-# LANGUAGE GADTs #-}
module T2006 where
data Expr a vs where
EPrim :: String -> a -> Expr a vs
EVar :: Expr a (a,vs)
interpret :: Expr a () -> a
interpret (EPrim _ a) = a
-- interpret EVar = error "unreachable"
| olsner/ghc | testsuite/tests/pmcheck/should_compile/T2006.hs | bsd-3-clause | 305 | 0 | 8 | 67 | 79 | 45 | 34 | 8 | 1 |
module ShouldCompile where
x@_ = x
| ryantm/ghc | testsuite/tests/deSugar/should_compile/ds-wildcard.hs | bsd-3-clause | 36 | 0 | 5 | 7 | 13 | 8 | 5 | 2 | 1 |
{-# LANGUAGE Trustworthy #-}
{-# LANGUAGE CPP, MagicHash, UnboxedTuples, NoImplicitPrelude #-}
{-# OPTIONS_HADDOCK hide #-}
-----------------------------------------------------------------------------
-- |
-- Module : GHC.Float.RealFracMethods
-- Copyright : (c) Daniel Fischer 2010
-- License : see libraries/base/LICENSE
--
-- Maintainer : [email protected]
-- Stability : internal
-- Portability : non-portable (GHC Extensions)
--
-- Methods for the RealFrac instances for 'Float' and 'Double',
-- with specialised versions for 'Int'.
--
-- Moved to their own module to not bloat GHC.Float further.
--
-----------------------------------------------------------------------------
#include "MachDeps.h"
module GHC.Float.RealFracMethods
( -- * Double methods
-- ** Integer results
properFractionDoubleInteger
, truncateDoubleInteger
, floorDoubleInteger
, ceilingDoubleInteger
, roundDoubleInteger
-- ** Int results
, properFractionDoubleInt
, floorDoubleInt
, ceilingDoubleInt
, roundDoubleInt
-- * Double/Int conversions, wrapped primops
, double2Int
, int2Double
-- * Float methods
-- ** Integer results
, properFractionFloatInteger
, truncateFloatInteger
, floorFloatInteger
, ceilingFloatInteger
, roundFloatInteger
-- ** Int results
, properFractionFloatInt
, floorFloatInt
, ceilingFloatInt
, roundFloatInt
-- * Float/Int conversions, wrapped primops
, float2Int
, int2Float
) where
import GHC.Integer
import GHC.Base
import GHC.Num ()
#if WORD_SIZE_IN_BITS < 64
import GHC.IntWord64
#define TO64 integerToInt64
#define FROM64 int64ToInteger
#define MINUS64 minusInt64#
#define NEGATE64 negateInt64#
#else
#define TO64 integerToInt
#define FROM64 smallInteger
#define MINUS64 ( -# )
#define NEGATE64 negateInt#
uncheckedIShiftRA64# :: Int# -> Int# -> Int#
uncheckedIShiftRA64# = uncheckedIShiftRA#
uncheckedIShiftL64# :: Int# -> Int# -> Int#
uncheckedIShiftL64# = uncheckedIShiftL#
#endif
default ()
------------------------------------------------------------------------------
-- Float Methods --
------------------------------------------------------------------------------
-- Special Functions for Int, nice, easy and fast.
-- They should be small enough to be inlined automatically.
-- We have to test for ±0.0 to avoid returning -0.0 in the second
-- component of the pair. Unfortunately the branching costs a lot
-- of performance.
properFractionFloatInt :: Float -> (Int, Float)
properFractionFloatInt (F# x) =
if isTrue# (x `eqFloat#` 0.0#)
then (I# 0#, F# 0.0#)
else case float2Int# x of
n -> (I# n, F# (x `minusFloat#` int2Float# n))
-- truncateFloatInt = float2Int
floorFloatInt :: Float -> Int
floorFloatInt (F# x) =
case float2Int# x of
n | isTrue# (x `ltFloat#` int2Float# n) -> I# (n -# 1#)
| otherwise -> I# n
ceilingFloatInt :: Float -> Int
ceilingFloatInt (F# x) =
case float2Int# x of
n | isTrue# (int2Float# n `ltFloat#` x) -> I# (n +# 1#)
| otherwise -> I# n
roundFloatInt :: Float -> Int
roundFloatInt x = float2Int (c_rintFloat x)
-- Functions with Integer results
-- With the new code generator in GHC 7, the explicit bit-fiddling is
-- slower than the old code for values of small modulus, but when the
-- 'Int' range is left, the bit-fiddling quickly wins big, so we use that.
-- If the methods are called on smallish values, hopefully people go
-- through Int and not larger types.
-- Note: For negative exponents, we must check the validity of the shift
-- distance for the right shifts of the mantissa.
{-# INLINE properFractionFloatInteger #-}
properFractionFloatInteger :: Float -> (Integer, Float)
properFractionFloatInteger v@(F# x) =
case decodeFloat_Int# x of
(# m, e #)
| isTrue# (e <# 0#) ->
case negateInt# e of
s | isTrue# (s ># 23#) -> (0, v)
| isTrue# (m <# 0#) ->
case negateInt# (negateInt# m `uncheckedIShiftRA#` s) of
k -> (smallInteger k,
case m -# (k `uncheckedIShiftL#` s) of
r -> F# (encodeFloatInteger (smallInteger r) e))
| otherwise ->
case m `uncheckedIShiftRL#` s of
k -> (smallInteger k,
case m -# (k `uncheckedIShiftL#` s) of
r -> F# (encodeFloatInteger (smallInteger r) e))
| otherwise -> (shiftLInteger (smallInteger m) e, F# 0.0#)
{-# INLINE truncateFloatInteger #-}
truncateFloatInteger :: Float -> Integer
truncateFloatInteger x =
case properFractionFloatInteger x of
(n, _) -> n
-- floor is easier for negative numbers than truncate, so this gets its
-- own implementation, it's a little faster.
{-# INLINE floorFloatInteger #-}
floorFloatInteger :: Float -> Integer
floorFloatInteger (F# x) =
case decodeFloat_Int# x of
(# m, e #)
| isTrue# (e <# 0#) ->
case negateInt# e of
s | isTrue# (s ># 23#) -> if isTrue# (m <# 0#) then (-1) else 0
| otherwise -> smallInteger (m `uncheckedIShiftRA#` s)
| otherwise -> shiftLInteger (smallInteger m) e
-- ceiling x = -floor (-x)
-- If giving this its own implementation is faster at all,
-- it's only marginally so, hence we keep it short.
{-# INLINE ceilingFloatInteger #-}
ceilingFloatInteger :: Float -> Integer
ceilingFloatInteger (F# x) =
negateInteger (floorFloatInteger (F# (negateFloat# x)))
{-# INLINE roundFloatInteger #-}
roundFloatInteger :: Float -> Integer
roundFloatInteger x = float2Integer (c_rintFloat x)
------------------------------------------------------------------------------
-- Double Methods --
------------------------------------------------------------------------------
-- Special Functions for Int, nice, easy and fast.
-- They should be small enough to be inlined automatically.
-- We have to test for ±0.0 to avoid returning -0.0 in the second
-- component of the pair. Unfortunately the branching costs a lot
-- of performance.
properFractionDoubleInt :: Double -> (Int, Double)
properFractionDoubleInt (D# x) =
if isTrue# (x ==## 0.0##)
then (I# 0#, D# 0.0##)
else case double2Int# x of
n -> (I# n, D# (x -## int2Double# n))
-- truncateDoubleInt = double2Int
floorDoubleInt :: Double -> Int
floorDoubleInt (D# x) =
case double2Int# x of
n | isTrue# (x <## int2Double# n) -> I# (n -# 1#)
| otherwise -> I# n
ceilingDoubleInt :: Double -> Int
ceilingDoubleInt (D# x) =
case double2Int# x of
n | isTrue# (int2Double# n <## x) -> I# (n +# 1#)
| otherwise -> I# n
roundDoubleInt :: Double -> Int
roundDoubleInt x = double2Int (c_rintDouble x)
-- Functions with Integer results
-- The new Code generator isn't quite as good for the old 'Double' code
-- as for the 'Float' code, so for 'Double' the bit-fiddling also wins
-- when the values have small modulus.
-- When the exponent is negative, all mantissae have less than 64 bits
-- and the right shifting of sized types is much faster than that of
-- 'Integer's, especially when we can
-- Note: For negative exponents, we must check the validity of the shift
-- distance for the right shifts of the mantissa.
{-# INLINE properFractionDoubleInteger #-}
properFractionDoubleInteger :: Double -> (Integer, Double)
properFractionDoubleInteger v@(D# x) =
case decodeDoubleInteger x of
(# m, e #)
| isTrue# (e <# 0#) ->
case negateInt# e of
s | isTrue# (s ># 52#) -> (0, v)
| m < 0 ->
case TO64 (negateInteger m) of
n ->
case n `uncheckedIShiftRA64#` s of
k ->
(FROM64 (NEGATE64 k),
case MINUS64 n (k `uncheckedIShiftL64#` s) of
r ->
D# (encodeDoubleInteger (FROM64 (NEGATE64 r)) e))
| otherwise ->
case TO64 m of
n ->
case n `uncheckedIShiftRA64#` s of
k -> (FROM64 k,
case MINUS64 n (k `uncheckedIShiftL64#` s) of
r -> D# (encodeDoubleInteger (FROM64 r) e))
| otherwise -> (shiftLInteger m e, D# 0.0##)
{-# INLINE truncateDoubleInteger #-}
truncateDoubleInteger :: Double -> Integer
truncateDoubleInteger x =
case properFractionDoubleInteger x of
(n, _) -> n
-- floor is easier for negative numbers than truncate, so this gets its
-- own implementation, it's a little faster.
{-# INLINE floorDoubleInteger #-}
floorDoubleInteger :: Double -> Integer
floorDoubleInteger (D# x) =
case decodeDoubleInteger x of
(# m, e #)
| isTrue# (e <# 0#) ->
case negateInt# e of
s | isTrue# (s ># 52#) -> if m < 0 then (-1) else 0
| otherwise ->
case TO64 m of
n -> FROM64 (n `uncheckedIShiftRA64#` s)
| otherwise -> shiftLInteger m e
{-# INLINE ceilingDoubleInteger #-}
ceilingDoubleInteger :: Double -> Integer
ceilingDoubleInteger (D# x) =
negateInteger (floorDoubleInteger (D# (negateDouble# x)))
{-# INLINE roundDoubleInteger #-}
roundDoubleInteger :: Double -> Integer
roundDoubleInteger x = double2Integer (c_rintDouble x)
-- Wrappers around double2Int#, int2Double#, float2Int# and int2Float#,
-- we need them here, so we move them from GHC.Float and re-export them
-- explicitly from there.
double2Int :: Double -> Int
double2Int (D# x) = I# (double2Int# x)
int2Double :: Int -> Double
int2Double (I# i) = D# (int2Double# i)
float2Int :: Float -> Int
float2Int (F# x) = I# (float2Int# x)
int2Float :: Int -> Float
int2Float (I# i) = F# (int2Float# i)
-- Quicker conversions from 'Double' and 'Float' to 'Integer',
-- assuming the floating point value is integral.
--
-- Note: Since the value is integral, the exponent can't be less than
-- (-TYP_MANT_DIG), so we need not check the validity of the shift
-- distance for the right shfts here.
{-# INLINE double2Integer #-}
double2Integer :: Double -> Integer
double2Integer (D# x) =
case decodeDoubleInteger x of
(# m, e #)
| isTrue# (e <# 0#) ->
case TO64 m of
n -> FROM64 (n `uncheckedIShiftRA64#` negateInt# e)
| otherwise -> shiftLInteger m e
{-# INLINE float2Integer #-}
float2Integer :: Float -> Integer
float2Integer (F# x) =
case decodeFloat_Int# x of
(# m, e #)
| isTrue# (e <# 0#) -> smallInteger (m `uncheckedIShiftRA#` negateInt# e)
| otherwise -> shiftLInteger (smallInteger m) e
-- Foreign imports, the rounding is done faster in C when the value
-- isn't integral, so we call out for rounding. For values of large
-- modulus, calling out to C is slower than staying in Haskell, but
-- presumably 'round' is mostly called for values with smaller modulus,
-- when calling out to C is a major win.
-- For all other functions, calling out to C gives at most a marginal
-- speedup for values of small modulus and is much slower than staying
-- in Haskell for values of large modulus, so those are done in Haskell.
foreign import ccall unsafe "rintDouble"
c_rintDouble :: Double -> Double
foreign import ccall unsafe "rintFloat"
c_rintFloat :: Float -> Float
| tolysz/prepare-ghcjs | spec-lts8/base/GHC/Float/RealFracMethods.hs | bsd-3-clause | 11,762 | 0 | 29 | 3,141 | 2,251 | 1,197 | 1,054 | 186 | 2 |
-- | Geometric functions concerning angles. If not otherwise specified, all angles are in radians.
module Graphics.Gloss.Geometry.Angle
( degToRad
, radToDeg
, normaliseAngle )
where
-- | Convert degrees to radians
{-# INLINE degToRad #-}
degToRad :: Float -> Float
degToRad d = d * pi / 180
-- | Convert radians to degrees
{-# INLINE radToDeg #-}
radToDeg :: Float -> Float
radToDeg r = r * 180 / pi
-- | Normalise an angle to be between 0 and 2*pi radians
{-# INLINE normaliseAngle #-}
normaliseAngle :: Float -> Float
normaliseAngle f = f - 2 * pi * floor' (f / (2 * pi))
where floor' :: Float -> Float
floor' x = fromIntegral (floor x :: Int)
| gscalzo/HaskellTheHardWay | gloss-try/gloss-master/gloss/Graphics/Gloss/Geometry/Angle.hs | mit | 665 | 6 | 10 | 139 | 159 | 89 | 70 | 15 | 1 |
module Y2018.M02.D21.Solution where
{--
More P99 fun with lists from:
http://www.ic.unicamp.br/~meidanis/courses/mc336/2009s2/prolog/problemas/
P28 (**) Sorting a list of lists according to length of sublists
a) We suppose that a list (InList) contains elements that are lists themselves.
The objective is to sort the elements of InList according to their length. E.g.
short lists first, longer lists later, or vice versa.
Example:
?- lsort([[a,b,c],[d,e],[f,g,h],[d,e],[i,j,k,l],[m,n],[o]],L).
L = [[o], [d, e], [d, e], [m, n], [a, b, c], [f, g, h], [i, j, k, l]]
--}
import Control.Arrow ((&&&))
import Data.List (sortOn, groupBy)
import Data.Function (on)
lsort :: [[a]] -> [[a]]
lsort = sortOn length -- from Denis Stoyanov @xgrommx
-- create a list of lists by matching the lengths against the ints
-- until the ints or lengths is exhausted:
lengths, ints :: [Int]
lengths = [1,8,7,4,9,5,4,1,2,7,
7,6,4,9,4,5,6,5,5,5,
6,4,6,4,3,6,1,7,3,8,
1,8,2,6,1,1,3,5,8,7,
9,2,6,6,8,7,8,2,5,5]
ints = [33,61,14,26,50,22,87,61,98,24,
8,71,24,89,44,3,42,21,16,62,
8,94,82,57,59,18,50,30,54,47,
64,35,88,36,81,40,48,62,61,81,
31,37,13,99,35,29,33,94,81,29,
39,62,11,18,48,45,33,43,99,49,
17,49,21,44,33,16,33,48,73,35,
81,47,96,54,23,62,6,94,16,44,
95,4,76,14,71,2,4,70,22,13,
9,96,9,55,12,91,63,41,33,37]
-- random numbers provided by random.org
mklistolists :: [Int] -> [a] -> [[a]]
mklistolists [] _ = []
mklistolists _ [] = []
mklistolists (len:gths) list@(_:_) = let (subl,ist) = splitAt len list in
subl:mklistolists gths ist
-- what is your listolists? What is your lfsort listolists?
{--
>>> listies = mklistolists lengths ints
>>> length listies
20
>>> take 4 listies
[[33],[61,14,26,50,22,87,61,98],[24,8,71,24,89,44,3],[42,21,16,62]]
>>> lsort listies
[[33],[61],[37],[81,31],[42,21,16,62],[81,40,48,62],[49,21,44,33],[23,62,6,94],
[47,64,35,88,36],[16,44,95,4,76],[13,9,96,9,55],[12,91,63,41,33],
[45,33,43,99,49,17],[14,71,2,4,70,22],[24,8,71,24,89,44,3],
[37,13,99,35,29,33,94],[81,29,39,62,11,18,48],[61,14,26,50,22,87,61,98],
[8,94,82,57,59,18,50,30,54],[16,33,48,73,35,81,47,96,54]]
--}
{-- BONUS -----------------------------------------------------------------
b) Again, we suppose that a list (InList) contains elements that are lists
themselves. But this time the objective is to sort the elements of InList
according to their length frequency; i.e. in the default, where sorting is done
ascendingly, lists with rare lengths are placed first, others with a more
frequent length come later.
Example:
?- lfsort([[a,b,c],[d,e],[f,g,h],[d,e],[i,j,k,l],[m,n],[o]],L).
L = [[i, j, k, l], [o], [a, b, c], [f, g, h], [d, e], [d, e], [m, n]]
Note that in the above example, the first two lists in the result L have length
4 and 1, both lengths appear just once. The third and forth list have length 3
which appears, there are two list of this length. And finally, the last three
lists have length 2. This is the most frequent length.
--}
lfsort :: [[a]] -> [[a]]
lfsort = concat . lsort . groupBy ((==) `on` length) . lsort
{--
or from Bazzargh @bazzargh
let {g=flip on length;f=sortBy (g compare)} in (concat . f . groupBy (g (==)) . f)
>>> lfsort (mklistolists lengths ints)
[[81,31],[61,14,26,50,22,87,61,98],[45,33,43,99,49,17],[14,71,2,4,70,22],
[8,94,82,57,59,18,50,30,54],[16,33,48,73,35,81,47,96,54],[33],[61],[37],
[24,8,71,24,89,44,3],[37,13,99,35,29,33,94],[81,29,39,62,11,18,48],
[42,21,16,62],[81,40,48,62],[49,21,44,33],[23,62,6,94],[47,64,35,88,36],
[16,44,95,4,76],[13,9,96,9,55],[12,91,63,41,33]]
And there we go!
--}
| geophf/1HaskellADay | exercises/HAD/Y2018/M02/D21/Solution.hs | mit | 3,700 | 0 | 9 | 576 | 715 | 457 | 258 | 29 | 1 |
{-# OPTIONS_GHC -fno-warn-type-defaults -fno-warn-orphans -fno-warn-missing-fields #-}
{-# LANGUAGE StandaloneDeriving #-}
{-# LANGUAGE DeriveDataTypeable #-}
module Yage.Prelude
( module ClassyPrelude
, io, pass
, traceShowS, traceShowS', ioTime, printIOTime, traceWith, traceStack
, globFp
-- list functions
, zipWithTF
, offset0
, eqType, descending
, (<?), (?>)
, isLeft, isRight
, Identity()
, qStr
, module Text.Show
, module FilePath
, module DeepSeq
, module Default
, module Prelude
, module Proxy
, module Printf
) where
import qualified Prelude as Prelude
import ClassyPrelude
import Text.Printf as Printf (printf)
import Data.Typeable
import Data.Data
import Data.Proxy as Proxy
import Data.Traversable as Trav
import Data.Foldable as Fold
import Data.Functor.Identity ()
import Data.Default as Default
import Control.DeepSeq as DeepSeq
import Control.DeepSeq.Generics as DeepSeq
import Filesystem.Path.CurrentOS as FilePath (decodeString,
encodeString)
import Foreign.Ptr
import System.CPUTime
import System.FilePath.Glob
import Text.Printf
import Text.Show
import Language.Haskell.TH
import Language.Haskell.TH.Quote
import Debug.Trace (traceStack)
io :: (MonadIO m) => IO a -> m a
io = liftIO
pass :: IO ()
pass = return ()
traceShowS :: Show a => ShowS -> a -> a
traceShowS sf a = traceShow (sf $ show a) a
traceShowS' :: Show a => String -> a -> a
traceShowS' msg = traceShowS (msg Prelude.++)
traceWith :: Show b => (a -> b) -> a -> a
traceWith f a = traceShow (f a) a
-- | time a monadic action in seconds, the monadic value is strict evaluated
ioTime :: MonadIO m => m a -> m (a, Double)
ioTime action = do
start <- io $! getCPUTime
v <- action
end <- v `seq` io $! getCPUTime
let diff = (fromIntegral (end - start)) / (10^(12::Int))
return $! (v, diff)
printIOTime :: MonadIO m => m a -> m a
printIOTime f = do
(res, t) <- ioTime f
_ <- io $! printf "Computation time: %0.5f sec\n" t
return res
-- stolen from: Graphics-GLUtil-BufferObjects
-- |A zero-offset 'Ptr'.
offset0 :: Ptr a
offset0 = offsetPtr 0
-- |Produce a 'Ptr' value to be used as an offset of the given number
-- of bytes.
offsetPtr :: Int -> Ptr a
offsetPtr = wordPtrToPtr . fromIntegral
eqType :: (Typeable r, Typeable t) => Proxy r -> Proxy t -> Bool
eqType r t = (typeOf r) == (typeOf t)
(?>) :: a -> Maybe a -> a
l ?> mr = maybe l id mr
(<?) :: Maybe a -> a -> a
(<?) = flip (?>)
descending :: (a -> a -> Ordering) -> (a -> a -> Ordering)
descending cmp = flip cmp
isLeft :: Either a b -> Bool
isLeft (Left _) = True
isLeft _ = False
isRight :: Either a b -> Bool
isRight (Right _)= True
isRight _ = False
zipWithTF :: (Traversable t, Foldable f) => (a -> b -> c) -> t a -> f b -> t c
zipWithTF g t f = snd (Trav.mapAccumL map_one (Fold.toList f) t)
where map_one (x:xs) y = (xs, g y x)
map_one _ _ = error "Yage.Prelude.zipWithTF"
qStr :: QuasiQuoter
qStr = QuasiQuoter { quoteExp = stringE }
-- | utility function to glob with a 'Filesystem.Path.FilePath'
--
-- > globFp ( "foo" </> "bar" </> "*<->.jpg" )
-- > ["foo/bar/image01.jpg", "foo/bar/image02.jpg"]
globFp :: MonadIO m => FilePath -> m [FilePath]
globFp = io . fmap (map fpFromString) . glob . fpToString
{-# INLINE globFp #-}
deriving instance Data Zero
deriving instance Typeable Zero
deriving instance Data a => Data (Succ a)
deriving instance Typeable Succ
| MaxDaten/yage-contrib | src/Yage/Prelude.hs | mit | 3,837 | 0 | 14 | 1,100 | 1,156 | 637 | 519 | 100 | 2 |
{-# LANGUAGE TemplateHaskell #-}
module PeaCoq where
import Control.Lens (makeLenses)
import Data.IORef (IORef)
import Data.IntMap (IntMap)
import Snap (Snaplet)
import Snap.Snaplet.Session (SessionManager)
import System.IO
import System.Process (ProcessHandle)
type Handles = (Handle, Handle, Handle, ProcessHandle)
data SessionState
= SessionState
Bool -- True while the session is alive
Handles -- I/O handles
-- Global state must be used in thread-safe way
data GlobalState
= GlobalState
{ gNextSession :: Int -- number to assign to the next session
, gActiveSessions :: IntMap SessionState
, gCoqtop :: String -- the command to use to run coqtop
}
type PeaCoqGlobRef = (IORef GlobalState)
type PeaCoqHash = String
type PeaCoqSession = SessionManager
-- Each thread gets a separate copy of this, fields must be read-only
data PeaCoq
= PeaCoq
{ _lGlobRef :: Snaplet PeaCoqGlobRef
, _lHash :: Snaplet PeaCoqHash
, _lSession :: Snaplet PeaCoqSession
}
-- Fields are lenses to separate concerns, use "Handler PeaCoq <Lens> a"
makeLenses ''PeaCoq
| Ptival/peacoq-server | lib/PeaCoq.hs | mit | 1,192 | 0 | 9 | 299 | 203 | 124 | 79 | 28 | 0 |
{-# LANGUAGE CPP #-}
#if __GLASGOW_HASKELL__ >= 702
{-# LANGUAGE Trustworthy #-}
#endif
-- | Reexports "Test.Hspec" from a @Trustworthy@ module.
module TestHspecTrustworthy (module Test.Hspec) where
import Test.Hspec
| haskell-compat/base-compat | base-compat-batteries/test/TestHspecTrustworthy.hs | mit | 218 | 0 | 5 | 29 | 22 | 16 | 6 | 3 | 0 |
{-#LANGUAGE ScopedTypeVariables #-}
{-#LANGUAGe DataKinds #-}
{-#LANGUAGE DeriveGeneric #-}
{-#LANGUAGE DeriveAnyClass #-}
{-#LANGUAGE FlexibleContexts #-}
module Foreign.Storable.Generic.Internal.GStorableSpec where
-- Test tools
import Test.Hspec
import Test.QuickCheck
import GenericType
-- Tested modules
import Foreign.Storable.Generic.Internal
-- Additional data
import Foreign.Storable.Generic -- overlapping Storable
import Foreign.Storable.Generic.Instances
import Data.Int
import Data.Word
import GHC.Generics
import Foreign.Ptr (Ptr, plusPtr)
import Foreign.Marshal.Alloc (malloc, mallocBytes, free)
import Foreign.Marshal.Array (peekArray,pokeArray)
data TestData = TestData Int Int64 Int8 Int8
deriving (Show, Generic, GStorable, Eq)
instance Arbitrary TestData where
arbitrary = TestData <$> arbitrary <*> arbitrary <*> arbitrary <*> arbitrary
data TestData2 = TestData2 Int8 TestData Int32 Int64
deriving (Show, Generic, GStorable, Eq)
instance Arbitrary TestData2 where
arbitrary = TestData2 <$> arbitrary <*> arbitrary <*> arbitrary <*> arbitrary
data TestData3 = TestData3 Int64 TestData2 Int16 TestData Int8
deriving (Show, Generic, GStorable, Eq)
instance Arbitrary TestData3 where
arbitrary = TestData3 <$> arbitrary <*> arbitrary <*> arbitrary <*> arbitrary <*> arbitrary
sizeEquality a = do
gsizeOf a `shouldBe` internalSizeOf (from a)
alignmentEquality a = do
gsizeOf a `shouldBe` internalSizeOf (from a)
pokeEquality a = do
let size = gsizeOf a
off <- generate $ suchThat arbitrary (>=0)
ptr <- mallocBytes (off + size)
-- First poke
gpokeByteOff ptr off a
bytes1 <- peekArray (off+size) ptr :: IO [Word8]
internalPokeByteOff ptr off (from a)
bytes2 <- peekArray (off+size) ptr :: IO [Word8]
free ptr
bytes1 `shouldBe` bytes2
peekEquality (a :: t) = do
let size = gsizeOf a
off <- generate $ suchThat arbitrary (>=0)
ptr <- mallocBytes (off + size)
bytes <- generate $ ok_vector (off+size) :: IO [Word8]
-- Save random stuff to memory
pokeArray ptr bytes
-- Take a peek
v1 <- gpeekByteOff ptr off :: IO t
v2 <- internalPeekByteOff ptr off :: IO (Rep t p)
free ptr
v1 `shouldBe` to v2
peekAndPoke (a :: t)= do
ptr <- malloc :: IO (Ptr t)
gpokeByteOff ptr 0 a
(gpeekByteOff ptr 0) `shouldReturn` a
spec :: Spec
spec = do
describe "gsizeOf" $ do
it "is equal to: internalSizeOf (from a)" $ property $ do
test1 <- generate $ arbitrary :: IO TestData
test2 <- generate $ arbitrary :: IO TestData2
test3 <- generate $ arbitrary :: IO TestData3
sizeEquality test1
sizeEquality test2
sizeEquality test3
describe "galignment" $ do
it "is equal to: internalAlignment (from a)" $ property $ do
test1 <- generate $ arbitrary :: IO TestData
test2 <- generate $ arbitrary :: IO TestData2
test3 <- generate $ arbitrary :: IO TestData3
alignmentEquality test1
alignmentEquality test2
alignmentEquality test3
describe "gpokeByteOff" $ do
it "is equal to: internalPokeByteOff ptr off (from a)" $ property $ do
test1 <- generate $ arbitrary :: IO TestData
test2 <- generate $ arbitrary :: IO TestData2
test3 <- generate $ arbitrary :: IO TestData3
pokeEquality test1
pokeEquality test2
pokeEquality test3
describe "gpeekByteOff" $ do
it "is equal to: to <$> internalPeekByteOff ptr off" $ property $ do
test1 <- generate $ arbitrary :: IO TestData
test2 <- generate $ arbitrary :: IO TestData2
test3 <- generate $ arbitrary :: IO TestData3
peekEquality test1
peekEquality test2
peekEquality test3
describe "Other tests:" $ do
it "gpokeByteOff ptr 0 val >> gpeekByteOff ptr 0 == val" $ property $ do
test1 <- generate $ arbitrary :: IO TestData
test2 <- generate $ arbitrary :: IO TestData2
test3 <- generate $ arbitrary :: IO TestData3
peekAndPoke test1
peekAndPoke test2
peekAndPoke test3
| mkloczko/derive-storable | test/Spec/Foreign/Storable/Generic/Internal/GStorableSpec.hs | mit | 4,354 | 0 | 14 | 1,211 | 1,220 | 594 | 626 | 100 | 1 |
module Drifter
(
-- * Managing Migrations
resolveDependencyOrder
, changeSequence
, migrate
-- * Types
, Drifter(..)
, ChangeName(..)
, Change(..)
, Description
, Method
, DBConnection
) where
-------------------------------------------------------------------------------
import Data.List
-------------------------------------------------------------------------------
import Drifter.Graph
import Drifter.Types
-------------------------------------------------------------------------------
-- | This is a helper for the common case of where you just want
-- dependencies to run in list order. This will take the input list
-- and set their dependencies to run in the given sequence.
changeSequence :: [Change a] -> [Change a]
changeSequence [] = []
changeSequence (x:xs) = reverse $ snd $ foldl' go (x, [x]) xs
where
go :: (Change a, [Change a]) -> Change a -> (Change a, [Change a])
go (lastChange, xs') c =
let c' = c { changeDependencies = [changeName lastChange] }
in (c', c':xs')
| AndrewRademacher/drifter | src/Drifter.hs | mit | 1,097 | 0 | 14 | 243 | 239 | 139 | 100 | 21 | 1 |
{-# LANGUAGE PatternSynonyms, ForeignFunctionInterface, JavaScriptFFI #-}
module GHCJS.DOM.JSFFI.Generated.SVGZoomAndPan
(pattern SVG_ZOOMANDPAN_UNKNOWN, pattern SVG_ZOOMANDPAN_DISABLE,
pattern SVG_ZOOMANDPAN_MAGNIFY, js_setZoomAndPan, setZoomAndPan,
js_getZoomAndPan, getZoomAndPan, SVGZoomAndPan,
castToSVGZoomAndPan, gTypeSVGZoomAndPan)
where
import Prelude ((.), (==), (>>=), return, IO, Int, Float, Double, Bool(..), Maybe, maybe, fromIntegral, round, fmap, Show, Read, Eq, Ord)
import Data.Typeable (Typeable)
import GHCJS.Types (JSVal(..), JSString)
import GHCJS.Foreign (jsNull)
import GHCJS.Foreign.Callback (syncCallback, asyncCallback, syncCallback1, asyncCallback1, syncCallback2, asyncCallback2, OnBlocked(..))
import GHCJS.Marshal (ToJSVal(..), FromJSVal(..))
import GHCJS.Marshal.Pure (PToJSVal(..), PFromJSVal(..))
import Control.Monad.IO.Class (MonadIO(..))
import Data.Int (Int64)
import Data.Word (Word, Word64)
import GHCJS.DOM.Types
import Control.Applicative ((<$>))
import GHCJS.DOM.EventTargetClosures (EventName, unsafeEventName)
import GHCJS.DOM.JSFFI.Generated.Enums
pattern SVG_ZOOMANDPAN_UNKNOWN = 0
pattern SVG_ZOOMANDPAN_DISABLE = 1
pattern SVG_ZOOMANDPAN_MAGNIFY = 2
foreign import javascript unsafe "$1[\"zoomAndPan\"] = $2;"
js_setZoomAndPan :: SVGZoomAndPan -> Word -> IO ()
-- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGZoomAndPan.zoomAndPan Mozilla SVGZoomAndPan.zoomAndPan documentation>
setZoomAndPan :: (MonadIO m) => SVGZoomAndPan -> Word -> m ()
setZoomAndPan self val = liftIO (js_setZoomAndPan (self) val)
foreign import javascript unsafe "$1[\"zoomAndPan\"]"
js_getZoomAndPan :: SVGZoomAndPan -> IO Word
-- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGZoomAndPan.zoomAndPan Mozilla SVGZoomAndPan.zoomAndPan documentation>
getZoomAndPan :: (MonadIO m) => SVGZoomAndPan -> m Word
getZoomAndPan self = liftIO (js_getZoomAndPan (self)) | manyoo/ghcjs-dom | ghcjs-dom-jsffi/src/GHCJS/DOM/JSFFI/Generated/SVGZoomAndPan.hs | mit | 1,960 | 14 | 8 | 235 | 477 | 292 | 185 | 31 | 1 |
{-# LANGUAGE TypeFamilies #-}
module Agent.PingPong.Role.Ask where
import AgentSystem.Generic
import Agent.PingPong
import qualified Agent.PingPong.Simple.Ask as Ask
import Data.IORef
--------------------------------------------------------------------------------
data PingRole = PingRole
data PongRole = PongRole
--------------------------------------------------------------------------------
instance RoleName PingRole where roleName _ = "Ping"
instance AgentRole PingRole where
type RoleState PingRole = (IORef Integer, IORef SomeAgentRef)
type RoleResult PingRole = ()
type RoleSysArgs PingRole = ()
type RoleArgs PingRole = (Integer, IORef SomeAgentRef)
instance RoleName PongRole where roleName _ = "Pong"
instance AgentRole PongRole where
type RoleState PongRole = ()
type RoleResult PongRole = ()
type RoleSysArgs PongRole = ()
type RoleArgs PongRole = ()
--------------------------------------------------------------------------------
pingRoleDescriptor = genericRoleDescriptor PingRole
(const $ return . uncurry Ask.pingDescriptor)
pongRoleDescriptor = genericRoleDescriptor PongRole
(const . const $ return Ask.pongDescriptor)
--------------------------------------------------------------------------------
runPingPong nPings = do pongRef <- newIORef undefined
putStrLn "<< CreateAgentOfRole >> "
let pingC = CreateAgentOfRole pingRoleDescriptor
(return ()) (return (nPings, pongRef))
pongC = CreateAgentOfRole pongRoleDescriptor
(return ()) (return ())
ping <- createAgentRef pingC
pong <- createAgentRef pongC
pongRef `writeIORef` someAgentRef pong
putStrLn "Starting PING"
agentStart ping
putStrLn "Starting PONG"
agentStart pong
putStrLn "Waiting PING termination"
agentWaitTermination ping
| fehu/h-agents | test/Agent/PingPong/Role/Ask.hs | mit | 2,150 | 0 | 13 | 618 | 413 | 211 | 202 | 39 | 1 |
module JoScript.Util.Text (foldlM, readFloat, readInt) where
import Prelude (read)
import Protolude hiding (foldlM)
import qualified Data.Text as T
foldlM :: Monad m => (b -> Char -> m b) -> b -> Text -> m b
foldlM f init bsInit = impl (pure init) bsInit where
impl acc bs
| T.null bs = acc
| otherwise = impl (acc >>= \acc' -> f acc' (T.head bs)) (T.tail bs)
readInt :: Text -> Integer
readInt = read . T.unpack
readFloat :: Text -> Double
readFloat = read . T.unpack
| AKST/jo | source/lib/JoScript/Util/Text.hs | mit | 486 | 0 | 15 | 105 | 217 | 115 | 102 | 13 | 1 |